diff --git itests/src/test/resources/testconfiguration.properties itests/src/test/resources/testconfiguration.properties index e3a25e7fb3..abe52fe7c6 100644 --- itests/src/test/resources/testconfiguration.properties +++ itests/src/test/resources/testconfiguration.properties @@ -2234,7 +2234,278 @@ minillaplocal.query.files=\ groupby8.q,\ groupby8_map.q,\ groupby8_map_skew.q,\ - groupby8_noskew.q + groupby8_noskew.q,\ + groupby9.q,\ + groupby_complex_types.q,\ + groupby_complex_types_multi_single_reducer.q,\ + groupby_cube1.q,\ + groupby_cube_multi_gby.q,\ + groupby_distinct_samekey.q,\ + groupby_duplicate_key.q,\ + groupby_grouping_id3.q,\ + groupby_grouping_sets1.q,\ + groupby_grouping_sets2.q,\ + groupby_grouping_sets3.q,\ + groupby_grouping_sets4.q,\ + groupby_grouping_sets5.q,\ + groupby_grouping_sets6.q,\ + groupby_grouping_sets_grouping.q,\ + groupby_grouping_sets_limit.q,\ + groupby_grouping_sets_view.q,\ + groupby_grouping_window.q,\ + groupby_join_pushdown.q,\ + groupby_map_ppr.q,\ + groupby_map_ppr_multi_distinct.q,\ + groupby_multi_insert_common_distinct.q,\ + groupby_multi_single_reducer.q,\ + groupby_multi_single_reducer2.q,\ + groupby_multi_single_reducer3.q,\ + groupby_multialias.q,\ + groupby_nocolumnalign.q,\ + groupby_position.q,\ + groupby_ppd.q,\ + groupby_ppr.q,\ + groupby_ppr_multi_distinct.q,\ + groupby_rollup1.q,\ + groupby_sort_10.q,\ + groupby_sort_11.q,\ + groupby_sort_1_23.q,\ + groupby_sort_2.q,\ + groupby_sort_3.q,\ + groupby_sort_4.q,\ + groupby_sort_5.q,\ + groupby_sort_6.q,\ + groupby_sort_7.q,\ + groupby_sort_8.q,\ + groupby_sort_9.q,\ + groupby_sort_skew_1_23.q,\ + groupby_sort_test_1.q,\ + groupingset_high_columns.q,\ + hashjoin.q,\ + having2.q,\ + hll.q,\ + implicit_cast1.q,\ + implicit_cast_during_insert.q,\ + implicit_decimal.q,\ + in_typecheck_char.q,\ + in_typecheck_mixed.q,\ + in_typecheck_pointlook.q,\ + in_typecheck_varchar.q,\ + infer_bucket_sort_convert_join.q,\ + infer_bucket_sort_dyn_part.q,\ + infer_bucket_sort_grouping_operators.q,\ + infer_bucket_sort_list_bucket.q,\ + infer_bucket_sort_map_operators.q,\ + infer_bucket_sort_merge.q,\ + infer_bucket_sort_multi_insert.q,\ + infer_bucket_sort_num_buckets.q,\ + infer_bucket_sort_reducers_power_two.q,\ + infer_const_type.q,\ + infer_join_preds.q,\ + innerjoin.q,\ + innerjoin1.q,\ + input.q,\ + input0.q,\ + input11.q,\ + input11_limit.q,\ + input12.q,\ + input13.q,\ + input1_limit.q,\ + input21.q,\ + input22.q,\ + input23.q,\ + input25.q,\ + input26.q,\ + input2_limit.q,\ + input30.q,\ + input32.q,\ + input3_limit.q,\ + input4.q,\ + input42.q,\ + input4_limit.q,\ + input6.q,\ + input7.q,\ + input8.q,\ + input9.q,\ + input_columnarserde.q,\ + input_dynamicserde.q,\ + input_lazyserde.q,\ + input_lazyserde2.q,\ + input_limit.q,\ + input_part0.q,\ + input_part1.q,\ + input_part10.q,\ + input_part2.q,\ + input_part3.q,\ + input_part4.q,\ + input_part5.q,\ + input_part6.q,\ + input_part7.q,\ + input_part8.q,\ + input_part9.q,\ + input_testsequencefile.q,\ + input_testxpath.q,\ + input_testxpath2.q,\ + input_testxpath3.q,\ + input_testxpath4.q,\ + insert0.q,\ + insert2_overwrite_partitions.q,\ + insert_nonacid_from_acid.q,\ + insertoverwrite_bucket.q,\ + intersect_all_rj.q,\ + interval_3.q,\ + interval_alt.q,\ + interval_arithmetic.q,\ + ivyDownload.q,\ + join10.q,\ + join11.q,\ + join12.q,\ + join13.q,\ + join14.q,\ + join15.q,\ + join16.q,\ + join17.q,\ + join18.q,\ + join18_multi_distinct.q,\ + join19.q,\ + join2.q,\ + join20.q,\ + join21.q,\ + join22.q,\ + join23.q,\ + join25.q,\ + join26.q,\ + join27.q,\ + join28.q,\ + join29.q,\ + join3.q,\ + join30.q,\ + join31.q,\ + join32.q,\ + join33.q,\ + join34.q,\ + join35.q,\ + join36.q,\ + join37.q,\ + join38.q,\ + join39.q,\ + join4.q,\ + join40.q,\ + join41.q,\ + join42.q,\ + join43.q,\ + join44.q,\ + join45.q,\ + join47.q,\ + join5.q,\ + join6.q,\ + join7.q,\ + join8.q,\ + join9.q,\ + join_1to1.q,\ + join_alt_syntax.q,\ + join_by_range_rule_not_null.q,\ + join_cond_pushdown_1.q,\ + join_cond_pushdown_2.q,\ + join_cond_pushdown_3.q,\ + join_cond_pushdown_4.q,\ + join_cond_pushdown_unqual1.q,\ + join_cond_pushdown_unqual2.q,\ + join_cond_pushdown_unqual3.q,\ + join_cond_pushdown_unqual4.q,\ + join_cond_pushdown_unqual5.q,\ + join_filters_overlap.q,\ + join_grp_diff_keys.q,\ + join_hive_626.q,\ + join_merge_multi_expressions.q,\ + join_merging.q,\ + join_on_varchar.q,\ + join_parse.q,\ + join_rc.q,\ + join_reorder.q,\ + join_reorder2.q,\ + join_reorder3.q,\ + join_reorder4.q,\ + join_star.q,\ + join_thrift.q,\ + join_view.q,\ + keyword_1.q,\ + lateral_view_cp.q,\ + lateral_view_explode2.q,\ + lateral_view_noalias.q,\ + lateral_view_onview.q,\ + lateral_view_onview2.q,\ + lateral_view_outer.q,\ + lateral_view_ppd.q,\ + leadlag_queries.q,\ + limit_pushdown2.q,\ + limit_pushdown_negative.q,\ + lineage1.q,\ + list_bucket_dml_1.q,\ + list_bucket_dml_11.q,\ + list_bucket_dml_12.q,\ + list_bucket_dml_13.q,\ + list_bucket_dml_14.q,\ + list_bucket_dml_2.q,\ + list_bucket_dml_3.q,\ + list_bucket_dml_4.q,\ + list_bucket_dml_5.q,\ + list_bucket_dml_6.q,\ + list_bucket_dml_7.q,\ + list_bucket_dml_9.q,\ + list_bucket_query_multiskew_1.q,\ + list_bucket_query_multiskew_2.q,\ + list_bucket_query_multiskew_3.q,\ + list_bucket_query_oneskew_1.q,\ + list_bucket_query_oneskew_2.q,\ + list_bucket_query_oneskew_3.q,\ + literal_decimal.q,\ + literal_double.q,\ + literal_ints.q,\ + literal_string.q,\ + llap_reader.q,\ + load_dyn_part10.q,\ + load_dyn_part11.q,\ + load_dyn_part12.q,\ + load_dyn_part13.q,\ + load_dyn_part14.q,\ + load_dyn_part4.q,\ + load_dyn_part6.q,\ + load_dyn_part8.q,\ + load_dyn_part9.q,\ + load_static_ptn_into_bucketed_table.q,\ + louter_join_ppr.q,\ + macro.q,\ + mapreduce5.q,\ + mapreduce6.q,\ + masking_1.q,\ + masking_10.q,\ + masking_12.q,\ + masking_13.q,\ + masking_1_newdb.q,\ + masking_2.q,\ + masking_3.q,\ + masking_4.q,\ + masking_6.q,\ + masking_7.q,\ + masking_8.q,\ + masking_9.q,\ + masking_disablecbo_1.q,\ + masking_disablecbo_2.q,\ + masking_disablecbo_3.q,\ + masking_disablecbo_4.q,\ + masking_mv.q,\ + masking_reserved.q,\ + merge3.q,\ + merge4.q,\ + merge_dynamic_partition.q,\ + merge_dynamic_partition2.q,\ + merge_dynamic_partition3.q,\ + merge_dynamic_partition4.q,\ + merge_dynamic_partition5.q,\ + merge_empty.q,\ + merge_join_1.q,\ + mergejoins.q encrypted.query.files=encryption_join_unencrypted_tbl.q,\ encryption_insert_partition_static.q,\ diff --git ql/src/test/queries/clientpositive/groupby_duplicate_key.q ql/src/test/queries/clientpositive/groupby_duplicate_key.q index ce5a091785..a2e1faf12c 100644 --- ql/src/test/queries/clientpositive/groupby_duplicate_key.q +++ ql/src/test/queries/clientpositive/groupby_duplicate_key.q @@ -1,18 +1,18 @@ --! qt:dataset:src explain -select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows); +select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key; -select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows); +select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key; explain create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows); +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key; create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows); +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key; -select key,dummy1,dummy2 from dummy_n6; +select key,dummy1,dummy2 from dummy_n6 order by key; explain -select max('pants'), max('pANTS') from src group by key limit 1; -select max('pants'), max('pANTS') from src group by key limit 1; +select max('pants'), max('pANTS') from src group by key order by key limit 1; +select max('pants'), max('pANTS') from src group by key order by key limit 1; diff --git ql/src/test/queries/clientpositive/groupby_grouping_sets_limit.q ql/src/test/queries/clientpositive/groupby_grouping_sets_limit.q index cda9c039dc..3c4a973538 100644 --- ql/src/test/queries/clientpositive/groupby_grouping_sets_limit.q +++ ql/src/test/queries/clientpositive/groupby_grouping_sets_limit.q @@ -6,31 +6,31 @@ CREATE TABLE T1_n141(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS T LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n141; EXPLAIN -SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10; -SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10; EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10; -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10; EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10; -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10; +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10; EXPLAIN -SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10; +SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10; -SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10; +SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10; EXPLAIN -SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10; +SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10; -SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10; +SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10; EXPLAIN -SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10; +SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10; -SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10; +SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10; diff --git ql/src/test/queries/clientpositive/groupby_grouping_sets_view.q ql/src/test/queries/clientpositive/groupby_grouping_sets_view.q index 376d3d4dc4..0489df741b 100644 --- ql/src/test/queries/clientpositive/groupby_grouping_sets_view.q +++ ql/src/test/queries/clientpositive/groupby_grouping_sets_view.q @@ -1,3 +1,5 @@ +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; create database test; diff --git ql/src/test/queries/clientpositive/groupby_sort_9.q ql/src/test/queries/clientpositive/groupby_sort_9.q index 0f9be6a30d..7e4c75e3c6 100644 --- ql/src/test/queries/clientpositive/groupby_sort_9.q +++ ql/src/test/queries/clientpositive/groupby_sort_9.q @@ -1,3 +1,5 @@ +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; set hive.exec.reducers.max = 10; set hive.map.groupby.sorted=true; diff --git ql/src/test/queries/clientpositive/groupingset_high_columns.q ql/src/test/queries/clientpositive/groupingset_high_columns.q index 977ced6efa..7107348921 100644 --- ql/src/test/queries/clientpositive/groupingset_high_columns.q +++ ql/src/test/queries/clientpositive/groupingset_high_columns.q @@ -1,3 +1,5 @@ +-- SORT_QUERY_RESULTS + create table facts (val string); insert into facts values ('abcdefghijklmnopqrstuvwxyz0123456789'); diff --git ql/src/test/queries/clientpositive/infer_bucket_sort_grouping_operators.q ql/src/test/queries/clientpositive/infer_bucket_sort_grouping_operators.q index f3f22ae34e..9f09356374 100644 --- ql/src/test/queries/clientpositive/infer_bucket_sort_grouping_operators.q +++ ql/src/test/queries/clientpositive/infer_bucket_sort_grouping_operators.q @@ -1,5 +1,7 @@ --! qt:dataset:src --! qt:dataset:part +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; set hive.exec.infer.bucket.sort=true; diff --git ql/src/test/queries/clientpositive/insert0.q ql/src/test/queries/clientpositive/insert0.q index d87c6b8a4c..4d02a73e19 100644 --- ql/src/test/queries/clientpositive/insert0.q +++ ql/src/test/queries/clientpositive/insert0.q @@ -16,7 +16,7 @@ INSERT INTO TABLE insert_into1_n1 SELECT * from src ORDER BY key DESC LIMIT 10; select * from insert_into1_n1 order by key; -create table ctas_table as SELECT key, count(value) as foo from src GROUP BY key LIMIT 10; +CREATE TABLE ctas_table AS SELECT key, count(value) AS foo FROM src GROUP BY key ORDER BY key LIMIT 10; describe extended ctas_table; @@ -28,7 +28,7 @@ set hive.exec.dynamic.partition=true; create table ctas_part (key int, value string) partitioned by (modkey bigint); insert overwrite table ctas_part partition (modkey) -select key, value, ceil(key / 100) from src where key is not null limit 10; +select key, value, ceil(key / 100) from src where key is not null order by key limit 10; select * from ctas_part order by key; @@ -36,4 +36,4 @@ select * from ctas_part order by key; DROP TABLE insert_into1_n1; DROP TABLE ctas_table; -DROP TABLE ctas_part; \ No newline at end of file +DROP TABLE ctas_part; diff --git ql/src/test/queries/clientpositive/join45.q ql/src/test/queries/clientpositive/join45.q index c0c7717dbe..cacd80efa3 100644 --- ql/src/test/queries/clientpositive/join45.q +++ ql/src/test/queries/clientpositive/join45.q @@ -10,6 +10,7 @@ FROM src1 JOIN src ON (src1.key=src.key AND src1.value between 100 and 102 AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -17,6 +18,7 @@ FROM src1 JOIN src ON (src1.key=src.key AND src1.value between 100 and 102 AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with pred on multiple inputs and none @@ -24,11 +26,13 @@ EXPLAIN SELECT * FROM src1 JOIN src ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with pred on single inputs and none @@ -38,6 +42,7 @@ FROM src1 JOIN src ON (src1.value between 100 and 102 AND src.value between 100 and 102 AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -45,6 +50,7 @@ FROM src1 JOIN src ON (src1.value between 100 and 102 AND src.value between 100 and 102 AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Disjunction with pred on multiple inputs and single inputs @@ -54,6 +60,7 @@ FROM src1 JOIN src ON (src1.key=src.key OR src1.value between 100 and 102 OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -61,6 +68,7 @@ FROM src1 JOIN src ON (src1.key=src.key OR src1.value between 100 and 102 OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with multiple inputs on one side @@ -69,12 +77,14 @@ SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Disjunction with multiple inputs on one side @@ -83,12 +93,14 @@ SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Function with multiple inputs on one side @@ -96,11 +108,13 @@ EXPLAIN SELECT * FROM src1 JOIN src ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Chained 1 @@ -109,12 +123,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 2 @@ -123,12 +139,14 @@ SELECT * FROM src LEFT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src LEFT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 3 @@ -137,12 +155,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 4 @@ -151,12 +171,14 @@ SELECT * FROM src RIGHT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src RIGHT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 5 @@ -165,12 +187,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 6 @@ -179,12 +203,14 @@ SELECT * FROM src FULL OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src FULL OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Right outer join with multiple inner joins and mixed conditions @@ -195,6 +221,7 @@ RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean LIMIT 10; SELECT * @@ -203,4 +230,5 @@ RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean LIMIT 10; diff --git ql/src/test/queries/clientpositive/join47.q ql/src/test/queries/clientpositive/join47.q index b5345d8492..bd34c091f3 100644 --- ql/src/test/queries/clientpositive/join47.q +++ ql/src/test/queries/clientpositive/join47.q @@ -11,6 +11,7 @@ FROM src1 JOIN src ON (src1.key=src.key AND src1.value between 100 and 102 AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -18,6 +19,7 @@ FROM src1 JOIN src ON (src1.key=src.key AND src1.value between 100 and 102 AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with pred on multiple inputs and none @@ -25,11 +27,13 @@ EXPLAIN SELECT * FROM src1 JOIN src ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with pred on single inputs and none @@ -39,6 +43,7 @@ FROM src1 JOIN src ON (src1.value between 100 and 102 AND src.value between 100 and 102 AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -46,6 +51,7 @@ FROM src1 JOIN src ON (src1.value between 100 and 102 AND src.value between 100 and 102 AND true) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Disjunction with pred on multiple inputs and single inputs @@ -55,6 +61,7 @@ FROM src1 JOIN src ON (src1.key=src.key OR src1.value between 100 and 102 OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * @@ -62,6 +69,7 @@ FROM src1 JOIN src ON (src1.key=src.key OR src1.value between 100 and 102 OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Conjunction with multiple inputs on one side @@ -70,12 +78,14 @@ SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Disjunction with multiple inputs on one side @@ -84,12 +94,14 @@ SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON (src1.key+src.key >= 100 OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Function with multiple inputs on one side @@ -97,11 +109,13 @@ EXPLAIN SELECT * FROM src1 JOIN src ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; SELECT * FROM src1 JOIN src ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value LIMIT 10; -- Chained 1 @@ -110,12 +124,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 2 @@ -124,12 +140,14 @@ SELECT * FROM src LEFT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src LEFT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 3 @@ -138,12 +156,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 4 @@ -152,12 +172,14 @@ SELECT * FROM src RIGHT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src RIGHT OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 5 @@ -166,12 +188,14 @@ SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src JOIN src1 a ON (a.key+src.key >= 100) FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Chained 6 @@ -180,12 +204,14 @@ SELECT * FROM src FULL OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; SELECT * FROM src FULL OUTER JOIN src1 a ON (a.key = src.key) JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value LIMIT 10; -- Right outer join with multiple inner joins and mixed conditions @@ -196,6 +222,7 @@ RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean LIMIT 10; SELECT * @@ -204,4 +231,5 @@ RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean LIMIT 10; diff --git ql/src/test/queries/clientpositive/join_star.q ql/src/test/queries/clientpositive/join_star.q index 70cc8f7c07..5d2ef96433 100644 --- ql/src/test/queries/clientpositive/join_star.q +++ ql/src/test/queries/clientpositive/join_star.q @@ -1,3 +1,5 @@ +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; create table fact(m1 int, m2 int, d1 int, d2 int); create table dim1(f1 int, f2 int); diff --git ql/src/test/queries/clientpositive/leadlag_queries.q ql/src/test/queries/clientpositive/leadlag_queries.q index 50210fee5e..7c62145385 100644 --- ql/src/test/queries/clientpositive/leadlag_queries.q +++ ql/src/test/queries/clientpositive/leadlag_queries.q @@ -1,4 +1,6 @@ --! qt:dataset:part +-- SORT_QUERY_RESULTS + -- 1. testLeadUDAF select p_mfgr, p_retailprice, lead(p_retailprice) over (partition by p_mfgr order by p_name) as l1, diff --git ql/src/test/queries/clientpositive/load_dyn_part11.q ql/src/test/queries/clientpositive/load_dyn_part11.q index 1e61981d9f..62be4fd70c 100644 --- ql/src/test/queries/clientpositive/load_dyn_part11.q +++ ql/src/test/queries/clientpositive/load_dyn_part11.q @@ -1,4 +1,6 @@ --! qt:dataset:srcpart +-- SORT_QUERY_RESULTS + show partitions srcpart; diff --git ql/src/test/queries/clientpositive/masking_1.q ql/src/test/queries/clientpositive/masking_1.q index d8d069f8ef..781c62a722 100644 --- ql/src/test/queries/clientpositive/masking_1.q +++ ql/src/test/queries/clientpositive/masking_1.q @@ -1,5 +1,7 @@ --! qt:dataset:srcpart --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactoryForTest; diff --git ql/src/test/queries/clientpositive/masking_2.q ql/src/test/queries/clientpositive/masking_2.q index eb581b41c1..e56edef2f9 100644 --- ql/src/test/queries/clientpositive/masking_2.q +++ ql/src/test/queries/clientpositive/masking_2.q @@ -1,4 +1,6 @@ --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactoryForTest; diff --git ql/src/test/queries/clientpositive/masking_3.q ql/src/test/queries/clientpositive/masking_3.q index 2317b5049e..012ce8b4e3 100644 --- ql/src/test/queries/clientpositive/masking_3.q +++ ql/src/test/queries/clientpositive/masking_3.q @@ -1,5 +1,7 @@ --! qt:dataset:srcpart --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; set hive.security.authorization.manager=org.apache.hadoop.hive.ql.security.authorization.plugin.sqlstd.SQLStdHiveAuthorizerFactoryForTest; diff --git ql/src/test/queries/clientpositive/masking_disablecbo_1.q ql/src/test/queries/clientpositive/masking_disablecbo_1.q index 74771e561e..b4067c8e7e 100644 --- ql/src/test/queries/clientpositive/masking_disablecbo_1.q +++ ql/src/test/queries/clientpositive/masking_disablecbo_1.q @@ -1,5 +1,7 @@ --! qt:dataset:srcpart --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.cbo.enable=false; set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; diff --git ql/src/test/queries/clientpositive/masking_disablecbo_2.q ql/src/test/queries/clientpositive/masking_disablecbo_2.q index 7be997509a..0c28c501d6 100644 --- ql/src/test/queries/clientpositive/masking_disablecbo_2.q +++ ql/src/test/queries/clientpositive/masking_disablecbo_2.q @@ -1,4 +1,6 @@ --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.cbo.enable=false; set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; diff --git ql/src/test/queries/clientpositive/masking_disablecbo_3.q ql/src/test/queries/clientpositive/masking_disablecbo_3.q index d76edc1919..94e8db4c6a 100644 --- ql/src/test/queries/clientpositive/masking_disablecbo_3.q +++ ql/src/test/queries/clientpositive/masking_disablecbo_3.q @@ -1,5 +1,7 @@ --! qt:dataset:srcpart --! qt:dataset:src +-- SORT_QUERY_RESULTS + set hive.cbo.enable=false; set hive.mapred.mode=nonstrict; set hive.security.authorization.enabled=true; diff --git ql/src/test/results/clientpositive/groupby_complex_types_multi_single_reducer.q.out ql/src/test/results/clientpositive/groupby_complex_types_multi_single_reducer.q.out deleted file mode 100644 index dd2ea4a357..0000000000 --- ql/src/test/results/clientpositive/groupby_complex_types_multi_single_reducer.q.out +++ /dev/null @@ -1,266 +0,0 @@ -PREHOOK: query: CREATE TABLE DEST1_n47(key ARRAY, value BIGINT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST1_n47 -POSTHOOK: query: CREATE TABLE DEST1_n47(key ARRAY, value BIGINT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST1_n47 -PREHOOK: query: CREATE TABLE DEST2_n8(key MAP, value BIGINT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST2_n8 -POSTHOOK: query: CREATE TABLE DEST2_n8(key MAP, value BIGINT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST2_n8 -PREHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 -INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n47 -PREHOOK: Output: default@dest2_n8 -POSTHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 -INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n47 -POSTHOOK: Output: default@dest2_n8 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0 - Stage-5 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-5 - Stage-1 depends on stages: Stage-6 - Stage-7 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - keys: array(key) (type: array) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: array) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: array) - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: bigint) - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - keys: map(key:value) (type: map) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: array) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: array) - null sort order: z - sort order: + - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: array), VALUE._col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 19280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 19280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n47 - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n47 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: map) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: map) - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: map) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: map) - null sort order: z - sort order: + - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: map), VALUE._col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 9280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 9280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n8 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n8 - - Stage: Stage-7 - Stats Work - Basic Stats Work: - -PREHOOK: query: FROM SRC -INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 -INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n47 -PREHOOK: Output: default@dest2_n8 -POSTHOOK: query: FROM SRC -INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 -INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n47 -POSTHOOK: Output: default@dest2_n8 -POSTHOOK: Lineage: dest1_n47.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n47.value EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: dest2_n8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n8.value EXPRESSION [(src)src.null, ] -PREHOOK: query: SELECT DEST1_n47.* FROM DEST1_n47 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n47 -#### A masked pattern was here #### -POSTHOOK: query: SELECT DEST1_n47.* FROM DEST1_n47 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n47 -#### A masked pattern was here #### -["0"] 3 -["10"] 1 -["100"] 2 -["103"] 2 -["104"] 2 -["105"] 1 -["11"] 1 -["111"] 1 -["113"] 2 -["114"] 1 -PREHOOK: query: SELECT DEST2_n8.* FROM DEST2_n8 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n8 -#### A masked pattern was here #### -POSTHOOK: query: SELECT DEST2_n8.* FROM DEST2_n8 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n8 -#### A masked pattern was here #### -{"0":"val_0"} 3 -{"10":"val_10"} 1 -{"100":"val_100"} 2 -{"103":"val_103"} 2 -{"104":"val_104"} 2 -{"105":"val_105"} 1 -{"11":"val_11"} 1 -{"111":"val_111"} 1 -{"113":"val_113"} 2 -{"114":"val_114"} 1 diff --git ql/src/test/results/clientpositive/groupby_cube1.q.out ql/src/test/results/clientpositive/groupby_cube1.q.out deleted file mode 100644 index 0ac1490e34..0000000000 --- ql/src/test/results/clientpositive/groupby_cube1.q.out +++ /dev/null @@ -1,884 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n82(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n82 -POSTHOOK: query: CREATE TABLE T1_n82(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n82 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n82 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n82 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n82 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n82 -PREHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY CUBE(key, val) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY CUBE(key, val) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -1 11 1 -1 NULL 1 -2 12 1 -2 NULL 1 -3 13 1 -3 NULL 1 -7 17 1 -7 NULL 1 -8 18 1 -8 28 1 -8 NULL 2 -NULL 11 1 -NULL 12 1 -NULL 13 1 -NULL 17 1 -NULL 18 1 -NULL 28 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -1 11 0 1 -1 NULL 1 1 -2 12 0 1 -2 NULL 1 1 -3 13 0 1 -3 NULL 1 1 -7 17 0 1 -7 NULL 1 1 -8 18 0 1 -8 28 0 1 -8 NULL 1 2 -NULL 11 2 1 -NULL 12 2 1 -NULL 13 2 1 -NULL 17 2 1 -NULL 18 2 1 -NULL 28 2 1 -NULL NULL 3 6 -PREHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(DISTINCT val) - keys: key (type: string), 0L (type: bigint), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -1 11 1 -1 NULL 1 -2 12 1 -2 NULL 1 -3 13 1 -3 NULL 1 -7 17 1 -7 NULL 1 -8 18 1 -8 28 1 -8 NULL 2 -NULL 11 1 -NULL 12 1 -NULL 13 1 -NULL 17 1 -NULL 18 1 -NULL 28 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(DISTINCT val) - keys: key (type: string), 0L (type: bigint), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: final - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -NULL 6 -PREHOOK: query: CREATE TABLE T2_n51(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n51 -POSTHOOK: query: CREATE TABLE T2_n51(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n51 -PREHOOK: query: CREATE TABLE T3_n16(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n16 -POSTHOOK: query: CREATE TABLE T3_n16(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n16 -PREHOOK: query: EXPLAIN -FROM T1_n82 -INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube -INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -PREHOOK: Output: default@t2_n51 -PREHOOK: Output: default@t3_n16 -POSTHOOK: query: EXPLAIN -FROM T1_n82 -INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube -INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -POSTHOOK: Output: default@t2_n51 -POSTHOOK: Output: default@t3_n16 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0, Stage-5, Stage-9 - Stage-5 depends on stages: Stage-3 - Stage-8 depends on stages: Stage-1, Stage-5, Stage-9 - Stage-6 depends on stages: Stage-2 - Stage-7 depends on stages: Stage-6 - Stage-1 depends on stages: Stage-7 - Stage-9 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n82 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(1) - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(1) - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n51 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, val - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n51 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key1, key2, val - Column Types: string, string, int - Table: default.t2_n51 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-8 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key1, key2, val - Column Types: string, string, int - Table: default.t3_n16 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t3_n16 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, val - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t3_n16 - - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM T1_n82 -INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube -INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n82 -PREHOOK: Output: default@t2_n51 -PREHOOK: Output: default@t3_n16 -POSTHOOK: query: FROM T1_n82 -INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube -INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n82 -POSTHOOK: Output: default@t2_n51 -POSTHOOK: Output: default@t3_n16 -POSTHOOK: Lineage: t2_n51.key1 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n51.key2 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n51.val EXPRESSION [(t1_n82)t1_n82.null, ] -POSTHOOK: Lineage: t3_n16.key1 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t3_n16.key2 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: t3_n16.val EXPRESSION [(t1_n82)t1_n82.null, ] diff --git ql/src/test/results/clientpositive/groupby_cube_multi_gby.q.out ql/src/test/results/clientpositive/groupby_cube_multi_gby.q.out deleted file mode 100644 index af37eaca1a..0000000000 --- ql/src/test/results/clientpositive/groupby_cube_multi_gby.q.out +++ /dev/null @@ -1,250 +0,0 @@ -PREHOOK: query: create table t1_n21 like src -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1_n21 -POSTHOOK: query: create table t1_n21 like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1_n21 -PREHOOK: query: create table t2_n13 like src -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t2_n13 -POSTHOOK: query: create table t2_n13 like src -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t2_n13 -PREHOOK: query: explain from src -insert into table t1_n21 select -key, GROUPING__ID -group by cube(key, value) -insert into table t2_n13 select -key, value -group by key, value grouping sets ((key), (key, value)) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@t1_n21 -PREHOOK: Output: default@t2_n13 -POSTHOOK: query: explain from src -insert into table t1_n21 select -key, GROUPING__ID -group by cube(key, value) -insert into table t2_n13 select -key, value -group by key, value grouping sets ((key), (key, value)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@t1_n21 -POSTHOOK: Output: default@t2_n13 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), value (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), value (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n21 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n21 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.t1_n21 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.t2_n13 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n13 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n13 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - diff --git ql/src/test/results/clientpositive/groupby_duplicate_key.q.out ql/src/test/results/clientpositive/groupby_duplicate_key.q.out deleted file mode 100644 index 44e8ef6952..0000000000 --- ql/src/test/results/clientpositive/groupby_duplicate_key.q.out +++ /dev/null @@ -1,319 +0,0 @@ -PREHOOK: query: explain -select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Row Limit Per Split: 10 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), '' (type: string), '' (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), '' (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), '' (type: string) - Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), '' (type: string), '' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), '' (type: string), '' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -165 -238 -255 -27 -278 -311 -409 -484 -86 -98 -PREHOOK: query: explain -create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@dummy_n6 -POSTHOOK: query: explain -create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dummy_n6 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-0, Stage-3 - Stage-2 depends on stages: Stage-4 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Row Limit Per Split: 10 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), 'X' (type: string), 'X' (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), 'X' (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), 'X' (type: string) - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), 'X' (type: string), 'X' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), 'X' (type: string), 'X' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dummy_n6 - Select Operator - expressions: _col0 (type: string), 'X' (type: string), 'X' (type: string) - outputColumnNames: col1, col2, col3 - Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-4 - Create Table - columns: key string, dummy1 string, dummy2 string - name: default.dummy_n6 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat - serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, dummy1, dummy2 - Column Types: string, string, string - Table: default.dummy_n6 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@dummy_n6 -POSTHOOK: query: create table dummy_n6 as -select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dummy_n6 -POSTHOOK: Lineage: dummy_n6.dummy1 SIMPLE [] -POSTHOOK: Lineage: dummy_n6.dummy2 SIMPLE [] -POSTHOOK: Lineage: dummy_n6.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select key,dummy1,dummy2 from dummy_n6 -PREHOOK: type: QUERY -PREHOOK: Input: default@dummy_n6 -#### A masked pattern was here #### -POSTHOOK: query: select key,dummy1,dummy2 from dummy_n6 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dummy_n6 -#### A masked pattern was here #### -165 X X -238 X X -255 X X -27 X X -278 X X -311 X X -409 X X -484 X X -86 X X -98 X X -PREHOOK: query: explain -select max('pants'), max('pANTS') from src group by key limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select max('pants'), max('pANTS') from src group by key limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: max('pants'), max('pANTS') - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: string), _col2 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: max(VALUE._col0), max(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 92000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 1 - Processor Tree: - ListSink - -PREHOOK: query: select max('pants'), max('pANTS') from src group by key limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select max('pants'), max('pANTS') from src group by key limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -pants pANTS diff --git ql/src/test/results/clientpositive/groupby_grouping_id3.q.out ql/src/test/results/clientpositive/groupby_grouping_id3.q.out deleted file mode 100644 index cdc063b370..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_id3.q.out +++ /dev/null @@ -1,207 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n86(key INT, value INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n86 -POSTHOOK: query: CREATE TABLE T1_n86(key INT, value INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n86 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n86 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n86 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n86 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n86 -PREHOOK: query: EXPLAIN -SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n86 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (_col2 = 1L) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1L (type: bigint) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: int), KEY._col1 (type: int), 1L (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -1 NULL 1 2 -2 NULL 1 1 -3 NULL 1 2 -4 NULL 1 1 -PREHOOK: query: EXPLAIN -SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n86 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (_col2 = 1L) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, value, GROUPING__ID, count(*) -FROM T1_n86 -GROUP BY key, value -GROUPING SETS ((), (key)) -HAVING GROUPING__ID = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n86 -#### A masked pattern was here #### -1 NULL 1 2 -2 NULL 1 1 -3 NULL 1 2 -4 NULL 1 1 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets1.q.out ql/src/test/results/clientpositive/groupby_grouping_sets1.q.out deleted file mode 100644 index 43ab99b9f1..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets1.q.out +++ /dev/null @@ -1,620 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n41(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n41 -POSTHOOK: query: CREATE TABLE T1_n41(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n41 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n41 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n41 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n41 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n41 -PREHOOK: query: SELECT * FROM T1_n41 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM T1_n41 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -t1_n41.a t1_n41.b t1_n41.c -1 1 3 -2 2 4 -2 3 5 -3 2 8 -5 2 2 -8 1 1 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n41 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n41 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) from T1_n41 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) from T1_n41 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a b _c2 -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n41 group by cube(a, b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n41 group by cube(a, b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) from T1_n41 group by cube(a, b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) from T1_n41 group by cube(a, b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a b _c2 -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a b _c2 -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a b _c2 -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -PREHOOK: query: EXPLAIN -SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string), c (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) - null sort order: zzzz - sort order: ++++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) - Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a -1 -2 -3 -5 -8 -NULL -NULL -NULL -NULL -NULL -NULL -NULL -NULL -NULL -PREHOOK: query: EXPLAIN -SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -a -1 -2 -3 -5 -8 -PREHOOK: query: EXPLAIN -SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n41 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: (UDFToDouble(a) + UDFToDouble(b)) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n41 -#### A masked pattern was here #### -_c0 _c1 -2.0 1 -4.0 1 -5.0 2 -7.0 1 -9.0 1 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets2.q.out ql/src/test/results/clientpositive/groupby_grouping_sets2.q.out deleted file mode 100644 index 7831a49e95..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets2.q.out +++ /dev/null @@ -1,479 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n81(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n81 -POSTHOOK: query: CREATE TABLE T1_n81(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n81 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n81 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n81 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n81 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n81 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n81 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n81 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n81 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n81 group by cube(a, b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n81 group by cube(a, b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n81 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) from T1_n81 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) from T1_n81 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -a b _c2 -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT a, b, sum(c) from T1_n81 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, sum(c) from T1_n81 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n81 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(c) - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: double) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: double) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, sum(c) from T1_n81 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, sum(c) from T1_n81 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -#### A masked pattern was here #### -a b _c2 -1 1 3.0 -1 NULL 3.0 -2 2 4.0 -2 3 5.0 -2 NULL 9.0 -3 2 8.0 -3 NULL 8.0 -5 2 2.0 -5 NULL 2.0 -8 1 1.0 -8 NULL 1.0 -NULL 1 4.0 -NULL 2 14.0 -NULL 3 5.0 -NULL NULL 23.0 -PREHOOK: query: CREATE TABLE T2_n50(a STRING, b STRING, c int, d int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n50 -POSTHOOK: query: CREATE TABLE T2_n50(a STRING, b STRING, c int, d int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n50 -PREHOOK: query: INSERT OVERWRITE TABLE T2_n50 -SELECT a, b, c, c from T1_n81 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n81 -PREHOOK: Output: default@t2_n50 -POSTHOOK: query: INSERT OVERWRITE TABLE T2_n50 -SELECT a, b, c, c from T1_n81 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n81 -POSTHOOK: Output: default@t2_n50 -POSTHOOK: Lineage: t2_n50.a SIMPLE [(t1_n81)t1_n81.FieldSchema(name:a, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n50.b SIMPLE [(t1_n81)t1_n81.FieldSchema(name:b, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n50.c EXPRESSION [(t1_n81)t1_n81.FieldSchema(name:c, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n50.d EXPRESSION [(t1_n81)t1_n81.FieldSchema(name:c, type:string, comment:null), ] -_col0 _col1 _col2 _col3 -PREHOOK: query: EXPLAIN -SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n50 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n50 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n50 - Statistics: Num rows: 6 Data size: 1068 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: a (type: string), b (type: string), (c + d) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1068 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n50 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n50 -#### A masked pattern was here #### -a b _c2 -1 1 6 -1 NULL 6 -2 2 8 -2 3 10 -2 NULL 18 -3 2 16 -3 NULL 16 -5 2 4 -5 NULL 4 -8 1 2 -8 NULL 2 -NULL 1 8 -NULL 2 28 -NULL 3 10 -NULL NULL 46 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets3.q.out ql/src/test/results/clientpositive/groupby_grouping_sets3.q.out deleted file mode 100644 index a08dd02490..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets3.q.out +++ /dev/null @@ -1,298 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n118(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n118 -POSTHOOK: query: CREATE TABLE T1_n118(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n118 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets1.txt' INTO TABLE T1_n118 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n118 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets1.txt' INTO TABLE T1_n118 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n118 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets2.txt' INTO TABLE T1_n118 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n118 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets2.txt' INTO TABLE T1_n118 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n118 -PREHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n118 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(c), count(c), count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by cube(a, b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by cube(a, b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n118 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(c), count(c), count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -a b _c2 _c3 -1 1 3.0 2 -1 2 2.0 1 -1 NULL 2.6666666666666665 3 -2 2 5.333333333333333 3 -2 3 5.0 2 -2 NULL 5.2 5 -3 2 8.0 1 -3 NULL 8.0 1 -5 1 2.0 1 -5 NULL 2.0 1 -8 1 1.0 2 -8 NULL 1.0 2 -NULL 1 2.0 5 -NULL 2 5.2 5 -NULL 3 5.0 2 -NULL NULL 3.8333333333333335 12 -PREHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -Explain -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n118 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(c), count(c), count() - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n118 -#### A masked pattern was here #### -a b _c2 _c3 -1 1 3.0 2 -1 2 2.0 1 -1 NULL 2.6666666666666665 3 -2 2 5.333333333333333 3 -2 3 5.0 2 -2 NULL 5.2 5 -3 2 8.0 1 -3 NULL 8.0 1 -5 1 2.0 1 -5 NULL 2.0 1 -8 1 1.0 2 -8 NULL 1.0 2 -NULL 1 2.0 5 -NULL 2 5.2 5 -NULL 3 5.0 2 -NULL NULL 3.8333333333333335 12 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out deleted file mode 100644 index b61aba926d..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets4.q.out +++ /dev/null @@ -1,603 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n143(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n143 -POSTHOOK: query: CREATE TABLE T1_n143(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n143 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n143 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n143 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n143 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n143 -PREHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq2 -on subq1.a = subq2.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq2 -on subq1.a = subq2.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -1 1 1 1 1 1 -1 1 1 1 NULL 1 -1 NULL 1 1 1 1 -1 NULL 1 1 NULL 1 -2 2 1 2 2 1 -2 2 1 2 3 1 -2 2 1 2 NULL 2 -2 3 1 2 2 1 -2 3 1 2 3 1 -2 3 1 2 NULL 2 -2 NULL 2 2 2 1 -2 NULL 2 2 3 1 -2 NULL 2 2 NULL 2 -PREHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2, Stage-5 - Stage-4 is a root stage - Stage-5 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n143 - filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) < 3.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 -join -(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 -on subq1.a = subq2.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n143 -#### A masked pattern was here #### -1 1 1 1 1 1 -1 1 1 1 NULL 1 -1 NULL 1 1 1 1 -1 NULL 1 1 NULL 1 -2 2 1 2 2 1 -2 2 1 2 3 1 -2 2 1 2 NULL 2 -2 3 1 2 2 1 -2 3 1 2 3 1 -2 3 1 2 NULL 2 -2 NULL 2 2 2 1 -2 NULL 2 2 3 1 -2 NULL 2 2 NULL 2 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets5.q.out ql/src/test/results/clientpositive/groupby_grouping_sets5.q.out deleted file mode 100644 index b6b4dcb339..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets5.q.out +++ /dev/null @@ -1,383 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n24(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n24 -POSTHOOK: query: CREATE TABLE T1_n24(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n24 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n24 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n24 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n24 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n24 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n24 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by cube(a, b) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by cube(a, b) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n24 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n24 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM -(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n24 -#### A masked pattern was here #### -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -8 NULL 1 -NULL 1 2 -NULL 2 3 -NULL 3 1 -NULL NULL 6 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets6.q.out ql/src/test/results/clientpositive/groupby_grouping_sets6.q.out deleted file mode 100644 index f6571b4645..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets6.q.out +++ /dev/null @@ -1,166 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n75(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n75 -POSTHOOK: query: CREATE TABLE T1_n75(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n75 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n75 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n75 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n75 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n75 -PREHOOK: query: EXPLAIN -SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n75 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) = 5.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -5 2 -5 NULL -PREHOOK: query: EXPLAIN -SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n75 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (UDFToDouble(a) = 5.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b FROM -(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res -WHERE res.a=5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n75 -#### A masked pattern was here #### -5 2 -5 NULL diff --git ql/src/test/results/clientpositive/groupby_grouping_sets_grouping.q.out ql/src/test/results/clientpositive/groupby_grouping_sets_grouping.q.out deleted file mode 100644 index 93e081b729..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets_grouping.q.out +++ /dev/null @@ -1,1408 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n64(key INT, value INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n64 -POSTHOOK: query: CREATE TABLE T1_n64(key INT, value INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n64 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n64 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n64 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n64 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n64 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 0 -1 NULL 0 0 0 -1 NULL 1 0 1 -2 2 0 0 0 -2 NULL 1 0 1 -3 3 0 0 0 -3 NULL 0 0 0 -3 NULL 1 0 1 -4 5 0 0 0 -4 NULL 1 0 1 -NULL NULL 3 1 1 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 0 -1 NULL 0 0 0 -1 NULL 1 0 1 -2 2 0 0 0 -2 NULL 1 0 1 -3 3 0 0 0 -3 NULL 0 0 0 -3 NULL 1 0 1 -4 5 0 0 0 -4 NULL 1 0 1 -NULL 1 2 1 0 -NULL 2 2 1 0 -NULL 3 2 1 0 -NULL 5 2 1 0 -NULL NULL 2 1 0 -NULL NULL 3 1 1 -PREHOOK: query: explain -select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (grouping(_col2, 1L) = 1L) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -NULL 1 -NULL 2 -NULL 3 -NULL 5 -NULL NULL -NULL NULL -PREHOOK: query: explain -select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((grouping(_col2, 1L) = 1L) or (grouping(_col2, 0L) = 1L)) (type: boolean) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1L) + grouping(_col2, 0L)) (type: bigint), CASE WHEN (((grouping(_col2, 1L) + grouping(_col2, 0L)) = 1L)) THEN (_col0) ELSE (null) END (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: bigint), _col3 (type: int) - null sort order: zz - sort order: -+ - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: int) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 NULL 1 -2 NULL 1 -3 NULL 1 -4 NULL 1 -NULL 1 1 -NULL 2 1 -NULL 3 1 -NULL 5 1 -NULL NULL 1 -NULL NULL 2 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 0 -1 NULL 0 0 0 -1 NULL 1 0 1 -2 2 0 0 0 -2 NULL 1 0 1 -3 3 0 0 0 -3 NULL 0 0 0 -3 NULL 1 0 1 -4 5 0 0 0 -4 NULL 1 0 1 -NULL NULL 3 1 1 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 0 -1 NULL 0 0 0 -1 NULL 1 0 1 -2 2 0 0 0 -2 NULL 1 0 1 -3 3 0 0 0 -3 NULL 0 0 0 -3 NULL 1 0 1 -4 5 0 0 0 -4 NULL 1 0 1 -NULL 1 2 1 0 -NULL 2 2 1 0 -NULL 3 2 1 0 -NULL 5 2 1 0 -NULL NULL 2 1 0 -NULL NULL 3 1 1 -PREHOOK: query: explain -select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (grouping(_col2, 1L) = 1) (type: boolean) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -NULL 1 -NULL 2 -NULL 3 -NULL 5 -NULL NULL -NULL NULL -PREHOOK: query: explain -select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((grouping(_col2, 1L) = 1) or (grouping(_col2, 0L) = 1)) (type: boolean) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1L) + grouping(_col2, 0L)) (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: bigint), CASE WHEN ((_col2 = 1L)) THEN (_col0) END (type: int) - null sort order: zz - sort order: -+ - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: int) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, grouping(key)+grouping(value) as x -from T1_n64 -group by cube(key, value) -having grouping(key) = 1 OR grouping(value) = 1 -order by x desc, case when x = 1 then key end -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 NULL 1 -2 NULL 1 -3 NULL 1 -4 NULL 1 -NULL 1 1 -NULL 2 1 -NULL 3 1 -NULL 5 1 -NULL NULL 1 -NULL NULL 2 -PREHOOK: query: explain -select key, value, grouping(key), grouping(value) -from T1_n64 -group by key, value -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, grouping(key), grouping(value) -from T1_n64 -group by key, value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), 0L (type: bigint), 0L (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, grouping(key), grouping(value) -from T1_n64 -group by key, value -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, grouping(key), grouping(value) -from T1_n64 -group by key, value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 -1 NULL 0 0 -2 2 0 0 -3 3 0 0 -3 NULL 0 0 -4 5 0 0 -PREHOOK: query: explain -select key, value, grouping(value) -from T1_n64 -group by key, value -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, grouping(value) -from T1_n64 -group by key, value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), 0L (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, grouping(value) -from T1_n64 -group by key, value -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, grouping(value) -from T1_n64 -group by key, value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 -1 NULL 0 -2 2 0 -3 3 0 -3 NULL 0 -4 5 0 -PREHOOK: query: explain -select key, value -from T1_n64 -group by key, value -having grouping(key) = 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value -from T1_n64 -group by key, value -having grouping(key) = 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value -from T1_n64 -group by key, value -having grouping(key) = 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value -from T1_n64 -group by key, value -having grouping(key) = 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 -1 NULL -2 2 -3 3 -3 NULL -4 5 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 -1 NULL 0 0 -1 NULL 1 1 -2 2 0 0 -2 NULL 1 1 -3 3 0 0 -3 NULL 0 0 -3 NULL 1 1 -4 5 0 0 -4 NULL 1 1 -NULL 1 2 2 -NULL 2 2 2 -NULL 3 2 2 -NULL 5 2 2 -NULL NULL 2 2 -NULL NULL 3 3 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0L, 1L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by cube(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by cube(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 -1 NULL 0 0 -1 NULL 1 2 -2 2 0 0 -2 NULL 1 2 -3 3 0 0 -3 NULL 0 0 -3 NULL 1 2 -4 5 0 0 -4 NULL 1 2 -NULL 1 2 1 -NULL 2 2 1 -NULL 3 2 1 -NULL 5 2 1 -NULL NULL 2 1 -NULL NULL 3 3 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L, 0L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(key, value) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 -1 NULL 0 0 -1 NULL 1 1 -2 2 0 0 -2 NULL 1 1 -3 3 0 0 -3 NULL 0 0 -3 NULL 1 1 -4 5 0 0 -4 NULL 1 1 -NULL NULL 3 3 -PREHOOK: query: explain -select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n64 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: int) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: key (type: int), value (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0L, 1L) (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by rollup(key, value) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -POSTHOOK: query: select key, value, `grouping__id`, grouping(value, key) -from T1_n64 -group by rollup(key, value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n64 -#### A masked pattern was here #### -1 1 0 0 -1 NULL 0 0 -1 NULL 1 2 -2 2 0 0 -2 NULL 1 2 -3 3 0 0 -3 NULL 0 0 -3 NULL 1 2 -4 5 0 0 -4 NULL 1 2 -NULL NULL 3 3 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets_limit.q.out ql/src/test/results/clientpositive/groupby_grouping_sets_limit.q.out deleted file mode 100644 index b4aa6d1dd0..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_sets_limit.q.out +++ /dev/null @@ -1,518 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n141(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n141 -POSTHOOK: query: CREATE TABLE T1_n141(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n141 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n141 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n141 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n141 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n141 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) from T1_n141 group by a, b with cube LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -PREHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: a, b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: a (type: string), b (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -1 1 1 -1 NULL 1 -2 2 1 -2 3 1 -2 NULL 2 -3 2 1 -3 NULL 1 -5 2 1 -5 NULL 1 -8 1 1 -PREHOOK: query: EXPLAIN -SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string) - outputColumnNames: a, b, c - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string), b (type: string), c (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) - null sort order: zzzz - sort order: ++++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) - Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -1 -2 -3 -5 -8 -NULL -NULL -NULL -NULL -NULL -PREHOOK: query: EXPLAIN -SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: a (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -1 -2 -3 -5 -8 -PREHOOK: query: EXPLAIN -SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n141 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: (UDFToDouble(a) + UDFToDouble(b)) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a+b) LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n141 -#### A masked pattern was here #### -2.0 1 -4.0 1 -5.0 2 -7.0 1 -9.0 1 diff --git ql/src/test/results/clientpositive/groupby_grouping_window.q.out ql/src/test/results/clientpositive/groupby_grouping_window.q.out deleted file mode 100644 index 21d92567d5..0000000000 --- ql/src/test/results/clientpositive/groupby_grouping_window.q.out +++ /dev/null @@ -1,210 +0,0 @@ -PREHOOK: query: create table t_n33(category int, live int, comments int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t_n33 -POSTHOOK: query: create table t_n33(category int, live int, comments int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t_n33 -PREHOOK: query: insert into table t_n33 select key, 0, 2 from src tablesample(3 rows) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@t_n33 -POSTHOOK: query: insert into table t_n33 select key, 0, 2 from src tablesample(3 rows) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@t_n33 -POSTHOOK: Lineage: t_n33.category EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: t_n33.comments SIMPLE [] -POSTHOOK: Lineage: t_n33.live SIMPLE [] -PREHOOK: query: explain -select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 -FROM t_n33 -GROUP BY category -GROUPING SETS ((), (category)) -HAVING max(comments) > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@t_n33 -#### A masked pattern was here #### -POSTHOOK: query: explain -select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 -FROM t_n33 -GROUP BY category -GROUPING SETS ((), (category)) -HAVING max(comments) > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_n33 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t_n33 - Statistics: Num rows: 3 Data size: 36 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: category (type: int), live (type: int), comments (type: int) - outputColumnNames: category, live, comments - Statistics: Num rows: 3 Data size: 36 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: max(live), max(comments) - keys: category (type: int), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: bigint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: bigint) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int), _col3 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: max(VALUE._col0), max(VALUE._col1) - keys: KEY._col0 (type: int), KEY._col1 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col2, _col3 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Filter Operator - predicate: (_col3 > 0) (type: boolean) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col3 (type: int) - null sort order: az - sort order: ++ - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey1 (type: int) - outputColumnNames: _col0, _col2, _col3 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - PTF Operator - Function definitions: - Input definition - input alias: ptf_0 - output shape: _col0: int, _col2: int, _col3: int - type: WINDOWING - Windowing table definition - input alias: ptf_1 - name: windowingtablefunction - order by: _col3 ASC NULLS LAST - partition by: _col0 - raw input shape: - window functions: - window function definition - alias: rank_window_0 - arguments: _col3 - name: rank - window function: GenericUDAFRankEvaluator - window frame: ROWS PRECEDING(MAX)~FOLLOWING(MAX) - isPivotResult: true - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), rank_window_0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 -FROM t_n33 -GROUP BY category -GROUPING SETS ((), (category)) -HAVING max(comments) > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@t_n33 -#### A masked pattern was here #### -POSTHOOK: query: select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 -FROM t_n33 -GROUP BY category -GROUPING SETS ((), (category)) -HAVING max(comments) > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_n33 -#### A masked pattern was here #### -NULL 0 2 1 -86 0 2 1 -238 0 2 1 -311 0 2 1 -PREHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -GROUPING SETS ((), (category)) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_n33 -#### A masked pattern was here #### -POSTHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -GROUPING SETS ((), (category)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_n33 -#### A masked pattern was here #### -0 NULL -0 NULL -0 NULL -1 NULL -PREHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -PREHOOK: type: QUERY -PREHOOK: Input: default@t_n33 -#### A masked pattern was here #### -POSTHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_n33 -#### A masked pattern was here #### -0 0 -0 0 -0 NULL -PREHOOK: query: SELECT grouping(category), lag(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -PREHOOK: type: QUERY -PREHOOK: Input: default@t_n33 -#### A masked pattern was here #### -POSTHOOK: query: SELECT grouping(category), lag(live) over(partition by grouping(category)) -FROM t_n33 -GROUP BY category, live -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_n33 -#### A masked pattern was here #### -0 NULL -0 0 -0 0 diff --git ql/src/test/results/clientpositive/groupby_join_pushdown.q.out ql/src/test/results/clientpositive/groupby_join_pushdown.q.out deleted file mode 100644 index 2138eae171..0000000000 --- ql/src/test/results/clientpositive/groupby_join_pushdown.q.out +++ /dev/null @@ -1,1847 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT f.key, g.key, count(g.key) -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT f.key, g.key, count(g.key) -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string), (_col1 * _col3) (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT f.key, g.key -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT f.key, g.key -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT DISTINCT f.value, g.value -FROM src f JOIN src g ON(f.value = g.value) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT DISTINCT f.value, g.value -FROM src f JOIN src g ON(f.value = g.value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT f.key, g.key, COUNT(*) -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT f.key, g.key, COUNT(*) -FROM src f JOIN src g ON(f.key = g.key) -GROUP BY f.key, g.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string), (_col1 * _col3) (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT f.ctinyint, g.ctinyint, SUM(f.cbigint) -FROM alltypesorc f JOIN alltypesorc g ON(f.cint = g.cint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT f.ctinyint, g.ctinyint, SUM(f.cbigint) -FROM alltypesorc f JOIN alltypesorc g ON(f.cint = g.cint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: cint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 146784 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int), cbigint (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: tinyint), _col2 (type: bigint) - TableScan - alias: g - filterExpr: cint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: tinyint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: int) - 1 _col1 (type: int) - outputColumnNames: _col0, _col2, _col3 - Statistics: Num rows: 13785 Data size: 183376 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2) - keys: _col0 (type: tinyint), _col3 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint), _col1 (type: tinyint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) - Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN -SELECT f.cbigint, g.cbigint, MAX(f.cint) -FROM alltypesorc f JOIN alltypesorc g ON(f.cbigint = g.cbigint) -GROUP BY f.cbigint, g.cbigint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT f.cbigint, g.cbigint, MAX(f.cint) -FROM alltypesorc f JOIN alltypesorc g ON(f.cbigint = g.cbigint) -GROUP BY f.cbigint, g.cbigint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: cbigint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 110088 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cbigint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 82188 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: cint (type: int), cbigint (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 82188 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: max(_col0) - keys: _col1 (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: max(VALUE._col0) - keys: KEY._col0 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: bigint) - 1 _col0 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: bigint), _col2 (type: bigint), _col1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: cbigint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cbigint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: cbigint (type: bigint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: bigint) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT f.ctinyint, g.ctinyint, MIN(f.ctinyint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT f.ctinyint, g.ctinyint, MIN(f.ctinyint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: min(_col0) - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: tinyint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: min(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: tinyint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: tinyint), _col2 (type: tinyint), _col1 (type: tinyint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT MIN(f.cint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT MIN(f.cint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: min(_col1) - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: min(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col1 - Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT count(f.ctinyint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT count(f.ctinyint) -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 132 Data size: 2112 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (_col1 * _col3) (type: bigint) - outputColumnNames: _col0 - Statistics: Num rows: 132 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 132 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT count(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT count(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col1) - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 132 Data size: 2512 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (_col1 * _col3) (type: bigint), _col0 (type: tinyint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 132 Data size: 2512 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (_col1 * _col3) (type: bigint), _col0 (type: tinyint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT pk_alltypesorc_1 PRIMARY KEY (ctinyint) DISABLE RELY -PREHOOK: type: ALTERTABLE_ADDCONSTRAINT -POSTHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT pk_alltypesorc_1 PRIMARY KEY (ctinyint) DISABLE RELY -POSTHOOK: type: ALTERTABLE_ADDCONSTRAINT -PREHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: g - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1161499 Data size: 13900620 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: tinyint), _col2 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint), _col1 (type: tinyint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) - Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: bigint), _col0 (type: tinyint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 17161 Data size: 205752 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 17161 Data size: 205752 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: ALTER TABLE alltypesorc DROP CONSTRAINT pk_alltypesorc_1 -PREHOOK: type: ALTERTABLE_DROPCONSTRAINT -POSTHOOK: query: ALTER TABLE alltypesorc DROP CONSTRAINT pk_alltypesorc_1 -POSTHOOK: type: ALTERTABLE_DROPCONSTRAINT -PREHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT uk_alltypesorc_1 UNIQUE (ctinyint) DISABLE RELY -PREHOOK: type: ALTERTABLE_ADDCONSTRAINT -POSTHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT uk_alltypesorc_1 UNIQUE (ctinyint) DISABLE RELY -POSTHOOK: type: ALTERTABLE_ADDCONSTRAINT -PREHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -PREHOOK: type: QUERY -PREHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT sum(f.cint), f.ctinyint -FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) -GROUP BY f.ctinyint, g.ctinyint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@alltypesorc -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: f - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint), cint (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: g - filterExpr: ctinyint is not null (type: boolean) - Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ctinyint is not null (type: boolean) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ctinyint (type: tinyint) - outputColumnNames: _col0 - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: tinyint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: tinyint) - Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: tinyint) - 1 _col0 (type: tinyint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 647260 Data size: 7739232 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: tinyint), _col2 (type: tinyint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: tinyint), _col1 (type: tinyint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) - Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: bigint), _col0 (type: tinyint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 17161 Data size: 205688 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 17161 Data size: 205688 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/groupby_map_ppr.q.out ql/src/test/results/clientpositive/groupby_map_ppr.q.out deleted file mode 100644 index 621a80aae8..0000000000 --- ql/src/test/results/clientpositive/groupby_map_ppr.q.out +++ /dev/null @@ -1,405 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n144(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n144 -POSTHOOK: query: CREATE TABLE dest1_n144(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n144 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n144 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n144 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n144 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n144 -OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -GROUP BY SUBSTR(`key`, 1, 1) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT _col1), sum(_col1) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 93500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93500 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col3 (type: double) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:src] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 31916 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2 - columns.comments - columns.types string:int:string -#### A masked pattern was here #### - name default.dest1_n144 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n144 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) - outputColumnNames: key, c1, c2 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2 - columns.comments - columns.types string:int:string -#### A masked pattern was here #### - name default.dest1_n144 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n144 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, c1, c2 - Column Types: string, int, string - Table: default.dest1_n144 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n144 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n144 -POSTHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n144 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n144 -POSTHOOK: Lineage: dest1_n144.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n144.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n144.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n144.* FROM dest1_n144 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n144 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n144.* FROM dest1_n144 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n144 -#### A masked pattern was here #### -0 1 00.0 -1 71 132828.0 -2 69 251142.0 -3 62 364008.0 -4 74 4105526.0 -5 6 5794.0 -6 5 6796.0 -7 6 71470.0 -8 8 81524.0 -9 7 92094.0 diff --git ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out deleted file mode 100644 index 4db2458fb0..0000000000 --- ql/src/test/results/clientpositive/groupby_map_ppr_multi_distinct.q.out +++ /dev/null @@ -1,407 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n174(key STRING, c1 INT, c2 STRING, C3 INT, c4 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n174 -POSTHOOK: query: CREATE TABLE dest1_n174(key STRING, c1 INT, c2 STRING, C3 INT, c4 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n174 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n174 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n174 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n174 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n174 -OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2`, SUM(DISTINCT SUBSTR(`value`, 5)) AS `_o__c3`, COUNT(DISTINCT `value`) AS `_o__c4` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -GROUP BY SUBSTR(`key`, 1, 1) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT _col1), sum(_col1), sum(DISTINCT _col1), count(DISTINCT _col2) - keys: _col0 (type: string), _col1 (type: string), _col2 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1000 Data size: 294000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - numBuckets: -1 - sort order: +++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1000 Data size: 294000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col4 (type: double) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:src] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1_n174 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n174 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) - outputColumnNames: key, c1, c2, c3, c4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2152 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct,struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1_n174 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n174 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, c1, c2, c3, c4 - Column Types: string, int, string, int, int - Table: default.dest1_n174 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 2152 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct,struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct,struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct:struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n174 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n174 -POSTHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n174 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n174 -POSTHOOK: Lineage: dest1_n174.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n174.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n174.c3 EXPRESSION [(srcpart)src.null, ] -POSTHOOK: Lineage: dest1_n174.c4 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n174.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n174.* FROM dest1_n174 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n174 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n174.* FROM dest1_n174 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n174 -#### A masked pattern was here #### -0 1 00.0 0 1 -1 71 132828.0 10044 71 -2 69 251142.0 15780 69 -3 62 364008.0 20119 62 -4 74 4105526.0 30965 74 -5 6 5794.0 278 6 -6 5 6796.0 331 5 -7 6 71470.0 447 6 -8 8 81524.0 595 8 -9 7 92094.0 577 7 diff --git ql/src/test/results/clientpositive/groupby_multi_insert_common_distinct.q.out ql/src/test/results/clientpositive/groupby_multi_insert_common_distinct.q.out deleted file mode 100644 index 991f343394..0000000000 --- ql/src/test/results/clientpositive/groupby_multi_insert_common_distinct.q.out +++ /dev/null @@ -1,297 +0,0 @@ -PREHOOK: query: create table dest1_n99(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n99 -POSTHOOK: query: create table dest1_n99(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n99 -PREHOOK: query: create table dest2_n27(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest2_n27 -POSTHOOK: query: create table dest2_n27(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest2_n27 -PREHOOK: query: explain -from src -insert overwrite table dest1_n99 select key, count(distinct value) group by key -insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n99 -PREHOOK: Output: default@dest2_n27 -POSTHOOK: query: explain -from src -insert overwrite table dest1_n99 select key, count(distinct value) group by key -insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n99 -POSTHOOK: Output: default@dest2_n27 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT value) - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT value) - keys: (key + key) (type: double), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n99 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n99 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest1_n99 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest2_n27 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: double), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n27 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n27 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: from src -insert overwrite table dest1_n99 select key, count(distinct value) group by key -insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n99 -PREHOOK: Output: default@dest2_n27 -POSTHOOK: query: from src -insert overwrite table dest1_n99 select key, count(distinct value) group by key -insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n99 -POSTHOOK: Output: default@dest2_n27 -POSTHOOK: Lineage: dest1_n99.cnt EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n99.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n27.cnt EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n27.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from dest1_n99 where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n99 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest1_n99 where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n99 -#### A masked pattern was here #### -0 1 -2 1 -4 1 -5 1 -8 1 -9 1 -PREHOOK: query: select * from dest2_n27 where key < 20 order by key limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n27 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest2_n27 where key < 20 order by key limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n27 -#### A masked pattern was here #### -0 1 -10 1 -16 1 -18 1 -4 1 -8 1 diff --git ql/src/test/results/clientpositive/groupby_multi_single_reducer2.q.out ql/src/test/results/clientpositive/groupby_multi_single_reducer2.q.out deleted file mode 100644 index d151470d6c..0000000000 --- ql/src/test/results/clientpositive/groupby_multi_single_reducer2.q.out +++ /dev/null @@ -1,283 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_g2_n4(key STRING, c1 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_g2_n4 -POSTHOOK: query: CREATE TABLE dest_g2_n4(key STRING, c1 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_g2_n4 -PREHOOK: query: CREATE TABLE dest_g3_n0(key STRING, c1 INT, c2 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_g3_n0 -POSTHOOK: query: CREATE TABLE dest_g3_n0(key STRING, c1 INT, c2 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_g3_n0 -PREHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) -INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest_g2_n4 -PREHOOK: Output: default@dest_g3_n0 -POSTHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) -INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest_g2_n4 -POSTHOOK: Output: default@dest_g3_n0 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: ((substr(key, 1, 1) < 5) or (substr(key, 1, 1) >= 5)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((substr(key, 1, 1) < 5) or (substr(key, 1, 1) >= 5)) (type: boolean) - Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: substr(key, 1, 1) (type: string), key (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: substr(key, 1, 1) (type: string) - Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (KEY._col0 >= 5) (type: boolean) - Statistics: Num rows: 110 Data size: 19360 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g2_n4 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, c1 - Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (KEY._col0 < 5) (type: boolean) - Statistics: Num rows: 110 Data size: 19360 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 105 Data size: 10605 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g3_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) - outputColumnNames: key, c1, c2 - Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g2_n4 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, c1 - Column Types: string, int - Table: default.dest_g2_n4 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, c1, c2 - Column Types: string, int, int - Table: default.dest_g3_n0 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g3_n0 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM src -INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) -INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest_g2_n4 -PREHOOK: Output: default@dest_g3_n0 -POSTHOOK: query: FROM src -INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) -INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest_g2_n4 -POSTHOOK: Output: default@dest_g3_n0 -POSTHOOK: Lineage: dest_g2_n4.c1 EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: dest_g2_n4.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_g3_n0.c1 EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: dest_g3_n0.c2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_g3_n0.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT * FROM dest_g2_n4 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_g2_n4 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM dest_g2_n4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_g2_n4 -#### A masked pattern was here #### -5 6 -6 5 -7 6 -8 8 -9 7 -PREHOOK: query: SELECT * FROM dest_g3_n0 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_g3_n0 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM dest_g3_n0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_g3_n0 -#### A masked pattern was here #### -0 1 3 -1 71 115 -2 69 111 -3 62 99 -4 74 124 -PREHOOK: query: DROP TABLE dest_g2_n4 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@dest_g2_n4 -PREHOOK: Output: default@dest_g2_n4 -POSTHOOK: query: DROP TABLE dest_g2_n4 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@dest_g2_n4 -POSTHOOK: Output: default@dest_g2_n4 -PREHOOK: query: DROP TABLE dest_g3_n0 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@dest_g3_n0 -PREHOOK: Output: default@dest_g3_n0 -POSTHOOK: query: DROP TABLE dest_g3_n0 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@dest_g3_n0 -POSTHOOK: Output: default@dest_g3_n0 diff --git ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out deleted file mode 100644 index 4b4d57f2a0..0000000000 --- ql/src/test/results/clientpositive/groupby_multi_single_reducer3.q.out +++ /dev/null @@ -1,1080 +0,0 @@ -PREHOOK: query: create table e1_n1 (key string, count int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@e1_n1 -POSTHOOK: query: create table e1_n1 (key string, count int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@e1_n1 -PREHOOK: query: create table e2_n2 (key string, count int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: create table e2_n2 (key string, count int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@e2_n2 -PREHOOK: query: explain -from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: explain -from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Reduce Operator Tree: - Forward - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((VALUE._col0) IN ('val_100', 'val_200', 'val_300') and (KEY._col0) IN (100, 150, 200)) (type: boolean) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: ((VALUE._col0) IN ('val_400', 'val_500') and (KEY._col0) IN (400, 450)) (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e1_n1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e2_n2 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from e1_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from e1_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -100 2 -200 2 -PREHOOK: query: select * from e2_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -POSTHOOK: query: select * from e2_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -400 1 -PREHOOK: query: explain -from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: explain -from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: value (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: value (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((VALUE._col0 + VALUE._col0) = 200) or ((VALUE._col0 - 100) = 100) or ((VALUE._col0 = 300) and KEY._col0 is not null)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (((VALUE._col0 + VALUE._col0) = 400) or (((VALUE._col0 - 100) = 500) and KEY._col0 is not null)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e1_n1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e2_n2 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from e1_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from e1_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -val_100 2 -val_200 2 -PREHOOK: query: select * from e2_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -POSTHOOK: query: select * from e2_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -val_200 2 -PREHOOK: query: explain -from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: explain -from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Reduce Operator Tree: - Forward - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((VALUE._col0) IN ('val_100', 'val_200', 'val_300') and (KEY._col0) IN (100, 150, 200)) (type: boolean) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: ((VALUE._col0) IN ('val_400', 'val_500') and (KEY._col0) IN (400, 450)) (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e1_n1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e2_n2 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: from src -insert overwrite table e1_n1 -select key, count(*) -where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) -group by key -insert overwrite table e2_n2 -select key, count(*) -where src.value in ('val_400', 'val_500') AND key in (400, 450) -group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from e1_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from e1_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -100 2 -200 2 -PREHOOK: query: select * from e2_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -POSTHOOK: query: select * from e2_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -400 1 -PREHOOK: query: explain -from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: explain -from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: value (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: value (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((VALUE._col0 + VALUE._col0) = 200) or ((VALUE._col0 - 100) = 100) or ((VALUE._col0 = 300) and KEY._col0 is not null)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (((VALUE._col0 + VALUE._col0) = 400) or (((VALUE._col0 - 100) = 500) and KEY._col0 is not null)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, count - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e1_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e1_n1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, count - Column Types: string, int - Table: default.e2_n2 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.e2_n2 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@e1_n1 -PREHOOK: Output: default@e2_n2 -POSTHOOK: query: from src -insert overwrite table e1_n1 -select value, count(*) -where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL -group by value -insert overwrite table e2_n2 -select value, count(*) -where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL -group by value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@e1_n1 -POSTHOOK: Output: default@e2_n2 -POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] -POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from e1_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from e1_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e1_n1 -#### A masked pattern was here #### -val_100 2 -val_200 2 -PREHOOK: query: select * from e2_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -POSTHOOK: query: select * from e2_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@e2_n2 -#### A masked pattern was here #### -val_200 2 diff --git ql/src/test/results/clientpositive/groupby_multialias.q.out ql/src/test/results/clientpositive/groupby_multialias.q.out deleted file mode 100644 index 1a42ff23a7..0000000000 --- ql/src/test/results/clientpositive/groupby_multialias.q.out +++ /dev/null @@ -1,73 +0,0 @@ -PREHOOK: query: create table t1_n150 (a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t1_n150 -POSTHOOK: query: create table t1_n150 (a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t1_n150 -PREHOOK: query: explain -select t1_n150.a as a1, min(t1_n150.a) as a -from t1_n150 -group by t1_n150.a -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n150 -#### A masked pattern was here #### -POSTHOOK: query: explain -select t1_n150.a as a1, min(t1_n150.a) as a -from t1_n150 -group by t1_n150.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n150 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n150 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: int) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: min(a) - keys: a (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: min(VALUE._col0) - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/groupby_position.q.out ql/src/test/results/clientpositive/groupby_position.q.out deleted file mode 100644 index 17f02c9089..0000000000 --- ql/src/test/results/clientpositive/groupby_position.q.out +++ /dev/null @@ -1,1891 +0,0 @@ -PREHOOK: query: CREATE TABLE testTable1(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@testTable1 -POSTHOOK: query: CREATE TABLE testTable1(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@testTable1 -PREHOOK: query: CREATE TABLE testTable2(key INT, val1 STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@testTable2 -POSTHOOK: query: CREATE TABLE testTable2(key INT, val1 STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@testTable2 -PREHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@testtable1 -PREHOOK: Output: default@testtable2 -POSTHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@testtable1 -POSTHOOK: Output: default@testtable2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 20) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 20) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), value (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 83 Data size: 22576 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable1 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.testtable1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.testtable2 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 83 Data size: 22576 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 15438 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable2 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable2 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@testtable1 -PREHOOK: Output: default@testtable2 -POSTHOOK: query: FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@testtable1 -POSTHOOK: Output: default@testtable2 -POSTHOOK: Lineage: testtable1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: testtable1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@testtable1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testtable1 -#### A masked pattern was here #### -0 1 -2 1 -4 1 -5 1 -8 1 -9 1 -10 1 -11 1 -12 1 -15 1 -17 1 -18 1 -19 1 -PREHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@testtable2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testtable2 -#### A masked pattern was here #### -0 val_0 1 -2 val_2 1 -4 val_4 1 -5 val_5 1 -8 val_8 1 -9 val_9 1 -10 val_10 1 -11 val_11 1 -12 val_12 1 -15 val_15 1 -17 val_17 1 -18 val_18 1 -19 val_19 1 -PREHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@testtable1 -PREHOOK: Output: default@testtable2 -POSTHOOK: query: EXPLAIN -FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@testtable1 -POSTHOOK: Output: default@testtable2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 20) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 20) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: value (type: string), key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 83 Data size: 22576 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable1 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.testtable1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.testtable2 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 83 Data size: 22576 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 15438 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable2 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 83 Data size: 23157 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.testtable2 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@testtable1 -PREHOOK: Output: default@testtable2 -POSTHOOK: query: FROM SRC -INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 -INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@testtable1 -POSTHOOK: Output: default@testtable2 -POSTHOOK: Lineage: testtable1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: testtable1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: testtable2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@testtable1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testtable1 -#### A masked pattern was here #### -0 1 -2 1 -4 1 -5 1 -8 1 -9 1 -10 1 -11 1 -12 1 -15 1 -17 1 -18 1 -19 1 -PREHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@testtable2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testtable2 -#### A masked pattern was here #### -0 val_0 1 -2 val_2 1 -4 val_4 1 -5 val_5 1 -8 val_8 1 -9 val_9 1 -10 val_10 1 -11 val_11 1 -12 val_12 1 -15 val_15 1 -17 val_17 1 -18 val_18 1 -19 val_19 1 -PREHOOK: query: EXPLAIN -SELECT t.key, t.value -FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t -ORDER BY 2 DESC, 1 ASC -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT t.key, t.value -FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t -ORDER BY 2 DESC, 1 ASC -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: b - filterExpr: (UDFToDouble(key) <= 20.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) <= 20.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: bigint), _col0 (type: string) - null sort order: zz - sort order: -+ - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT t.key, t.value -FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t -ORDER BY 2 DESC, 1 ASC -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT t.key, t.value -FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t -ORDER BY 2 DESC, 1 ASC -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 3 -5 3 -12 2 -15 2 -18 2 -10 1 -11 1 -17 1 -19 1 -2 1 -20 1 -4 1 -8 1 -9 1 -PREHOOK: query: EXPLAIN -SELECT c1, c2, c3, c4 -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 - ) a - JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT c1, c2, c3, c4 -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 - ) a - JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-3 depends on stages: Stage-2 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27 Data size: 9612 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - null sort order: zzzz - sort order: --++ - Statistics: Num rows: 27 Data size: 9612 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27 Data size: 9612 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 9612 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT c1, c2, c3, c4 -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 - ) a - JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT c1, c2, c3, c4 -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 - ) a - JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -19 val_19 19 val_19 -18 val_18 18 val_18 -17 val_17 17 val_17 -PREHOOK: query: EXPLAIN -SELECT key FROM src ORDER BY 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key FROM src ORDER BY 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key FROM src ORDER BY 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT key FROM src ORDER BY 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 -0 -0 -10 -100 -100 -103 -103 -104 -104 -105 -11 -111 -113 -113 -114 -116 -118 -118 -119 -119 -119 -12 -12 -120 -120 -125 -125 -126 -128 -128 -128 -129 -129 -131 -133 -134 -134 -136 -137 -137 -138 -138 -138 -138 -143 -145 -146 -146 -149 -149 -15 -15 -150 -152 -152 -153 -155 -156 -157 -158 -160 -162 -163 -164 -164 -165 -165 -166 -167 -167 -167 -168 -169 -169 -169 -169 -17 -170 -172 -172 -174 -174 -175 -175 -176 -176 -177 -178 -179 -179 -18 -18 -180 -181 -183 -186 -187 -187 -187 -189 -19 -190 -191 -191 -192 -193 -193 -193 -194 -195 -195 -196 -197 -197 -199 -199 -199 -2 -20 -200 -200 -201 -202 -203 -203 -205 -205 -207 -207 -208 -208 -208 -209 -209 -213 -213 -214 -216 -216 -217 -217 -218 -219 -219 -221 -221 -222 -223 -223 -224 -224 -226 -228 -229 -229 -230 -230 -230 -230 -230 -233 -233 -235 -237 -237 -238 -238 -239 -239 -24 -24 -241 -242 -242 -244 -247 -248 -249 -252 -255 -255 -256 -256 -257 -258 -26 -26 -260 -262 -263 -265 -265 -266 -27 -272 -272 -273 -273 -273 -274 -275 -277 -277 -277 -277 -278 -278 -28 -280 -280 -281 -281 -282 -282 -283 -284 -285 -286 -287 -288 -288 -289 -291 -292 -296 -298 -298 -298 -30 -302 -305 -306 -307 -307 -308 -309 -309 -310 -311 -311 -311 -315 -316 -316 -316 -317 -317 -318 -318 -318 -321 -321 -322 -322 -323 -325 -325 -327 -327 -327 -33 -331 -331 -332 -333 -333 -335 -336 -338 -339 -34 -341 -342 -342 -344 -344 -345 -348 -348 -348 -348 -348 -35 -35 -35 -351 -353 -353 -356 -360 -362 -364 -365 -366 -367 -367 -368 -369 -369 -369 -37 -37 -373 -374 -375 -377 -378 -379 -382 -382 -384 -384 -384 -386 -389 -392 -393 -394 -395 -395 -396 -396 -396 -397 -397 -399 -399 -4 -400 -401 -401 -401 -401 -401 -402 -403 -403 -403 -404 -404 -406 -406 -406 -406 -407 -409 -409 -409 -41 -411 -413 -413 -414 -414 -417 -417 -417 -418 -419 -42 -42 -421 -424 -424 -427 -429 -429 -43 -430 -430 -430 -431 -431 -431 -432 -435 -436 -437 -438 -438 -438 -439 -439 -44 -443 -444 -446 -448 -449 -452 -453 -454 -454 -454 -455 -457 -458 -458 -459 -459 -460 -462 -462 -463 -463 -466 -466 -466 -467 -468 -468 -468 -468 -469 -469 -469 -469 -469 -47 -470 -472 -475 -477 -478 -478 -479 -480 -480 -480 -481 -482 -483 -484 -485 -487 -489 -489 -489 -489 -490 -491 -492 -492 -493 -494 -495 -496 -497 -498 -498 -498 -5 -5 -5 -51 -51 -53 -54 -57 -58 -58 -64 -65 -66 -67 -67 -69 -70 -70 -70 -72 -72 -74 -76 -76 -77 -78 -8 -80 -82 -83 -83 -84 -84 -85 -86 -87 -9 -90 -90 -90 -92 -95 -95 -96 -97 -97 -98 -98 -PREHOOK: query: EXPLAIN -SELECT distinct key FROM src ORDER BY 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT distinct key FROM src ORDER BY 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT distinct key FROM src ORDER BY 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT distinct key FROM src ORDER BY 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 -10 -100 -103 -104 -105 -11 -111 -113 -114 -116 -118 -119 -12 -120 -125 -126 -128 -129 -131 -133 -134 -136 -137 -138 -143 -145 -146 -149 -15 -150 -152 -153 -155 -156 -157 -158 -160 -162 -163 -164 -165 -166 -167 -168 -169 -17 -170 -172 -174 -175 -176 -177 -178 -179 -18 -180 -181 -183 -186 -187 -189 -19 -190 -191 -192 -193 -194 -195 -196 -197 -199 -2 -20 -200 -201 -202 -203 -205 -207 -208 -209 -213 -214 -216 -217 -218 -219 -221 -222 -223 -224 -226 -228 -229 -230 -233 -235 -237 -238 -239 -24 -241 -242 -244 -247 -248 -249 -252 -255 -256 -257 -258 -26 -260 -262 -263 -265 -266 -27 -272 -273 -274 -275 -277 -278 -28 -280 -281 -282 -283 -284 -285 -286 -287 -288 -289 -291 -292 -296 -298 -30 -302 -305 -306 -307 -308 -309 -310 -311 -315 -316 -317 -318 -321 -322 -323 -325 -327 -33 -331 -332 -333 -335 -336 -338 -339 -34 -341 -342 -344 -345 -348 -35 -351 -353 -356 -360 -362 -364 -365 -366 -367 -368 -369 -37 -373 -374 -375 -377 -378 -379 -382 -384 -386 -389 -392 -393 -394 -395 -396 -397 -399 -4 -400 -401 -402 -403 -404 -406 -407 -409 -41 -411 -413 -414 -417 -418 -419 -42 -421 -424 -427 -429 -43 -430 -431 -432 -435 -436 -437 -438 -439 -44 -443 -444 -446 -448 -449 -452 -453 -454 -455 -457 -458 -459 -460 -462 -463 -466 -467 -468 -469 -47 -470 -472 -475 -477 -478 -479 -480 -481 -482 -483 -484 -485 -487 -489 -490 -491 -492 -493 -494 -495 -496 -497 -498 -5 -51 -53 -54 -57 -58 -64 -65 -66 -67 -69 -70 -72 -74 -76 -77 -78 -8 -80 -82 -83 -84 -85 -86 -87 -9 -90 -92 -95 -96 -97 -98 diff --git ql/src/test/results/clientpositive/groupby_ppd.q.out ql/src/test/results/clientpositive/groupby_ppd.q.out deleted file mode 100644 index 5731e9d5c2..0000000000 --- ql/src/test/results/clientpositive/groupby_ppd.q.out +++ /dev/null @@ -1,106 +0,0 @@ -PREHOOK: query: create table invites_n0 (id int, foo int, bar int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@invites_n0 -POSTHOOK: query: create table invites_n0 (id int, foo int, bar int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@invites_n0 -PREHOOK: query: explain select * from (select foo, bar from (select bar, foo from invites_n0 c union all select bar, foo from invites_n0 d) b) a group by bar, foo having bar=1 -PREHOOK: type: QUERY -PREHOOK: Input: default@invites_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from (select foo, bar from (select bar, foo from invites_n0 c union all select bar, foo from invites_n0 d) b) a group by bar, foo having bar=1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@invites_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: c - filterExpr: (bar = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (bar = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: foo (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: d - filterExpr: (bar = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (bar = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: foo (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), 1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: drop table invites_n0 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@invites_n0 -PREHOOK: Output: default@invites_n0 -POSTHOOK: query: drop table invites_n0 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@invites_n0 -POSTHOOK: Output: default@invites_n0 diff --git ql/src/test/results/clientpositive/groupby_ppr.q.out ql/src/test/results/clientpositive/groupby_ppr.q.out deleted file mode 100644 index bb5e7e65fa..0000000000 --- ql/src/test/results/clientpositive/groupby_ppr.q.out +++ /dev/null @@ -1,392 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n79(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n79 -POSTHOOK: query: CREATE TABLE dest1_n79(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n79 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n79 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n79 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1_n79 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n79 -OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -GROUP BY SUBSTR(`key`, 1, 1) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:src] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 31916 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2 - columns.comments - columns.types string:int:string -#### A masked pattern was here #### - name default.dest1_n79 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n79 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) - outputColumnNames: key, c1, c2 - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2 - columns.types string,int,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2 - columns.comments - columns.types string:int:string -#### A masked pattern was here #### - name default.dest1_n79 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n79 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, c1, c2 - Column Types: string, int, string - Table: default.dest1_n79 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: key (type: string), c1 (type: int), c2 (type: string) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2 - columns.types string,int,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2 - columns.types string,int,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll') - mode: complete - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n79 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n79 -POSTHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1_n79 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n79 -POSTHOOK: Lineage: dest1_n79.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n79.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n79.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n79.* FROM dest1_n79 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n79 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n79.* FROM dest1_n79 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n79 -#### A masked pattern was here #### -0 1 00.0 -1 71 132828.0 -2 69 251142.0 -3 62 364008.0 -4 74 4105526.0 -5 6 5794.0 -6 5 6796.0 -7 6 71470.0 -8 8 81524.0 -9 7 92094.0 diff --git ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out deleted file mode 100644 index 43f1ac82cd..0000000000 --- ql/src/test/results/clientpositive/groupby_ppr_multi_distinct.q.out +++ /dev/null @@ -1,780 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1(key STRING, c1 INT, c2 STRING, c3 INT, c4 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1 -POSTHOOK: query: CREATE TABLE dest1(key STRING, c1 INT, c2 STRING, c3 INT, c4 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1 -OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2`, SUM(DISTINCT SUBSTR(`value`, 5)) AS `_o__c3`, COUNT(DISTINCT `value`) AS `_o__c4` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -GROUP BY SUBSTR(`key`, 1, 1) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - numBuckets: -1 - sort order: +++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:src] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) - outputColumnNames: key, c1, c2, c3, c4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, c1, c2, c3, c4 - Column Types: string, int, string, int, int - Table: default.dest1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: key (type: string), c1 (type: int), c2 (type: string), c3 (type: int), c4 (type: int) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll'), compute_stats(VALUE._col4, 'hll'), compute_stats(VALUE._col5, 'hll') - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct:struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1 -POSTHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1 -POSTHOOK: Lineage: dest1.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.c3 EXPRESSION [(srcpart)src.null, ] -POSTHOOK: Lineage: dest1.c4 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1.* FROM dest1 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1.* FROM dest1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1 -#### A masked pattern was here #### -0 1 00.0 0 1 -1 71 132828.0 10044 71 -2 69 251142.0 15780 69 -3 62 364008.0 20119 62 -4 74 4105526.0 30965 74 -5 6 5794.0 278 6 -6 5 6796.0 331 5 -7 6 71470.0 447 6 -8 8 81524.0 595 8 -9 7 92094.0 577 7 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) - outputColumnNames: $f0, $f1, $f2 - Statistics: Num rows: 1000 Data size: 459000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: $f0 (type: string), $f1 (type: string), $f2 (type: string) - null sort order: zzz - numBuckets: -1 - sort order: +++ - Map-reduce partition columns: $f0 (type: string) - Statistics: Num rows: 1000 Data size: 459000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [src] - /srcpart/ds=2008-04-08/hr=12 [src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: $f0, $f1, $f2, $f3, $f4 - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: $f0 (type: string), UDFToInteger($f1) (type: int), concat($f0, $f2) (type: string), UDFToInteger($f3) (type: int), UDFToInteger($f4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1 - numFiles 1 - numRows 10 - rawDataSize 184 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 194 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) - outputColumnNames: key, c1, c2, c3, c4 - Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.comments - columns.types string:int:string:int:int -#### A masked pattern was here #### - name default.dest1 - numFiles 1 - numRows 10 - rawDataSize 184 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 194 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, c1, c2, c3, c4 - Column Types: string, int, string, int, int - Table: default.dest1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: key (type: string), c1 (type: int), c2 (type: string), c3 (type: int), c4 (type: int) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns key,c1,c2,c3,c4 - columns.types string,int,string,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll'), compute_stats(VALUE._col4, 'hll'), compute_stats(VALUE._col5, 'hll') - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct:struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1 -POSTHOOK: query: FROM srcpart src -INSERT OVERWRITE TABLE dest1 -SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) -WHERE src.ds = '2008-04-08' -GROUP BY substr(src.key,1,1) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1 -POSTHOOK: Lineage: dest1.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.c3 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1.c4 EXPRESSION [(srcpart)src.null, ] -POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1.* FROM dest1 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1.* FROM dest1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1 -#### A masked pattern was here #### -0 1 00.0 0 1 -1 71 132828.0 10044 71 -2 69 251142.0 15780 69 -3 62 364008.0 20119 62 -4 74 4105526.0 30965 74 -5 6 5794.0 278 6 -6 5 6796.0 331 5 -7 6 71470.0 447 6 -8 8 81524.0 595 8 -9 7 92094.0 577 7 diff --git ql/src/test/results/clientpositive/groupby_rollup1.q.out ql/src/test/results/clientpositive/groupby_rollup1.q.out deleted file mode 100644 index e7b61b4a33..0000000000 --- ql/src/test/results/clientpositive/groupby_rollup1.q.out +++ /dev/null @@ -1,717 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n91(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n91 -POSTHOOK: query: CREATE TABLE T1_n91(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n91 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n91 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n91 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n91 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n91 -PREHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n91 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -1 11 1 -1 NULL 1 -2 12 1 -2 NULL 1 -3 13 1 -3 NULL 1 -7 17 1 -7 NULL 1 -8 18 1 -8 28 1 -8 NULL 2 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n91 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(DISTINCT val) - keys: key (type: string), 0L (type: bigint), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n91 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -1 11 1 -1 NULL 1 -2 12 1 -2 NULL 1 -3 13 1 -3 NULL 1 -7 17 1 -7 NULL 1 -8 18 1 -8 28 1 -8 NULL 2 -NULL NULL 6 -PREHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n91 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(DISTINCT val) - keys: key (type: string), 0L (type: bigint), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: bigint) - mode: final - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -NULL 6 -PREHOOK: query: CREATE TABLE T2_n56(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n56 -POSTHOOK: query: CREATE TABLE T2_n56(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n56 -PREHOOK: query: CREATE TABLE T3_n20(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n20 -POSTHOOK: query: CREATE TABLE T3_n20(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n20 -PREHOOK: query: EXPLAIN -FROM T1_n91 -INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup -INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by rollup(key, val) -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -PREHOOK: Output: default@t2_n56 -PREHOOK: Output: default@t3_n20 -POSTHOOK: query: EXPLAIN -FROM T1_n91 -INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup -INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by rollup(key, val) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -POSTHOOK: Output: default@t2_n56 -POSTHOOK: Output: default@t3_n20 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0, Stage-5, Stage-9 - Stage-5 depends on stages: Stage-3 - Stage-8 depends on stages: Stage-1, Stage-5, Stage-9 - Stage-6 depends on stages: Stage-2 - Stage-7 depends on stages: Stage-6 - Stage-1 depends on stages: Stage-7 - Stage-9 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n91 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(1) - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(1) - keys: key (type: string), val (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n56 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n56 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key1, key2, val - Column Types: string, string, int - Table: default.t2_n56 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-8 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key1, key2, val - Column Types: string, string, int - Table: default.t3_n20 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: partials - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: final - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t3_n20 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, val - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t3_n20 - - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM T1_n91 -INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup -INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by key, val with rollup -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n91 -PREHOOK: Output: default@t2_n56 -PREHOOK: Output: default@t3_n20 -POSTHOOK: query: FROM T1_n91 -INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup -INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by key, val with rollup -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n91 -POSTHOOK: Output: default@t2_n56 -POSTHOOK: Output: default@t3_n20 -POSTHOOK: Lineage: t2_n56.key1 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n56.key2 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n56.val EXPRESSION [(t1_n91)t1_n91.null, ] -POSTHOOK: Lineage: t3_n20.key1 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t3_n20.key2 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: t3_n20.val EXPRESSION [(t1_n91)t1_n91.null, ] diff --git ql/src/test/results/clientpositive/groupby_sort_11.q.out ql/src/test/results/clientpositive/groupby_sort_11.q.out deleted file mode 100644 index 76d3c7c51a..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_11.q.out +++ /dev/null @@ -1,543 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n18(key STRING, val STRING) PARTITIONED BY (ds string) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n18 -POSTHOOK: query: CREATE TABLE T1_n18(key STRING, val STRING) PARTITIONED BY (ds string) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n18 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n18 PARTITION (ds='1') -SELECT * from src where key < 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@t1_n18@ds=1 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n18 PARTITION (ds='1') -SELECT * from src where key < 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@t1_n18@ds=1 -POSTHOOK: Lineage: t1_n18 PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: t1_n18 PARTITION(ds=1).val SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: EXPLAIN select count(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT key) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 465 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 5 Data size: 465 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -6 -PREHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: double) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0), count(VALUE._col1), count(VALUE._col2), sum(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -6 10 10 28.0 -PREHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint), _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col1), count(VALUE._col2), sum(DISTINCT KEY._col1:1._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 160 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 160 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -1 1 1 2.0 -1 1 1 4.0 -1 1 1 8.0 -1 1 1 9.0 -1 3 3 0.0 -1 3 3 5.0 -PREHOOK: query: EXPLAIN select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint), _col3 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col1), count(VALUE._col2), sum(DISTINCT KEY._col1:1._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -0 1 3 3 0.0 -2 1 1 1 2.0 -4 1 1 1 4.0 -5 1 3 3 5.0 -8 1 1 1 8.0 -9 1 1 1 9.0 -PREHOOK: query: EXPLAIN select count(distinct key+key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct key+key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: double) - mode: partial2 - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col0) - mode: partial2 - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct key+key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct key+key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -6 -PREHOOK: query: EXPLAIN select count(distinct 1) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct 1) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 10 Data size: 150 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT 1) - keys: 1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col0:0._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct 1) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct 1) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -1 -PREHOOK: query: EXPLAIN select count(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN select count(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n18 - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col0:0._col0) - mode: complete - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(distinct key) from T1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n18 -PREHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -POSTHOOK: query: select count(distinct key) from T1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n18 -POSTHOOK: Input: default@t1_n18@ds=1 -#### A masked pattern was here #### -6 diff --git ql/src/test/results/clientpositive/groupby_sort_1_23.q.out ql/src/test/results/clientpositive/groupby_sort_1_23.q.out deleted file mode 100644 index 8061ffad5e..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_1_23.q.out +++ /dev/null @@ -1,7809 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n80(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n80 -POSTHOOK: query: CREATE TABLE T1_n80(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n80 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n80 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n80 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n80 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n80 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n80 select key, val from T1_n80 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@t1_n80 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n80 select key, val from T1_n80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@t1_n80 -POSTHOOK: Lineage: t1_n80.key SIMPLE [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n80.val SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n18(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n18 -POSTHOOK: query: CREATE TABLE outputTbl1_n18(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n18 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T1_n80 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T1_n80 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T1_n80 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T1_n80 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: CREATE TABLE outputTbl2_n5(key1 int, key2 string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl2_n5 -POSTHOOK: query: CREATE TABLE outputTbl2_n5(key1 int, key2 string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl2_n5 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl2_n5 -SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl2_n5 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl2_n5 -SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl2_n5 -OPTIMIZED SQL: SELECT `key`, `val`, COUNT(*) AS `$f2` -FROM `default`.`t1_n80` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:string:int -#### A masked pattern was here #### - name default.outputtbl2_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n5 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:string:int -#### A masked pattern was here #### - name default.outputtbl2_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n5 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, string, int - Table: default.outputtbl2_n5 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n5 -SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl2_n5 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n5 -SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl2_n5 -POSTHOOK: Lineage: outputtbl2_n5.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl2_n5.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl2_n5.key2 SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl2_n5 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl2_n5 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl2_n5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl2_n5 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: CREATE TABLE outputTbl3_n2(key1 int, key2 int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl3_n2 -POSTHOOK: query: CREATE TABLE outputTbl3_n2(key1 int, key2 int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl3_n2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl3_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl3_n2 -OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, COUNT(*) AS `_o__c2` -FROM `default`.`t1_n80` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: 1 (type: int), UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, int, int - Table: default.outputtbl3_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - name: default.outputtbl3_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - name: default.outputtbl3_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl3_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl3_n2 -POSTHOOK: Lineage: outputtbl3_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl3_n2.key1 SIMPLE [] -POSTHOOK: Lineage: outputtbl3_n2.key2 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl3_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl3_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl3_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl3_n2 -#### A masked pattern was here #### -1 1 1 -1 2 1 -1 3 1 -1 7 1 -1 8 2 -PREHOOK: query: CREATE TABLE outputTbl4_n2(key1 int, key2 int, key3 string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl4_n2 -POSTHOOK: query: CREATE TABLE outputTbl4_n2(key1 int, key2 int, key3 string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl4_n2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl4_n2 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t1_n80` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl4_n2 -POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl3_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl3_n2 -OPTIMIZED SQL: SELECT `key` AS `$f0`, CAST(`key` AS DOUBLE) + 1 AS `$f1`, COUNT(*) AS `$f2` -FROM `default`.`t1_n80` -GROUP BY `key`, CAST(`key` AS DOUBLE) + 1 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: double) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0:t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n2 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, int, int - Table: default.outputtbl3_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl3_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 -SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl3_n2 -POSTHOOK: Lineage: outputtbl3_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl3_n2.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl3_n2.key2 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl3_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl3_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl3_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl3_n2 -#### A masked pattern was here #### -1 2 1 -2 3 1 -3 4 1 -7 8 1 -8 9 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key + key, sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -group by key + key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key + key, sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -group by key + key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS `$f0`, SUM(COUNT(*)) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0:t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key + key, sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -group by key + key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key + key, sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -group by key + key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -14 1 -16 2 -2 1 -4 1 -6 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n80 GROUP BY key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n80 GROUP BY key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -UNION ALL -SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0-subquery1:t1_n80, $hdt$_0-subquery2:t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n80 GROUP BY key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n80 GROUP BY key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -1 1 -2 1 -2 1 -3 1 -3 1 -7 1 -7 1 -8 2 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) FROM T1_n80 GROUP BY key + key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n80 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) FROM T1_n80 GROUP BY key + key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -GROUP BY `key` -UNION ALL -SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `key`, COUNT(*) AS `_o__c1` -FROM `default`.`t1_n80` -GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-2 depends on stages: Stage-9 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 - Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0-subquery2:$hdt$_0:t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: double) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: CAST( _col0 AS STRING) (type: string), _col1 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - TableScan - GatherStats: false - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0-subquery1:t1_n80] -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - name: default.outputtbl1_n18 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n80 GROUP BY key + key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT * FROM ( -SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n80 GROUP BY key + key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -14 1 -16 2 -2 1 -2 1 -3 1 -4 1 -6 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`$f1` + `t2`.`$f1` AS `_o__c1` -FROM (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t0` -INNER JOIN (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 0 - value expressions: _col1 (type: bigint) - auto parallelism: false - TableScan - alias: t1_n80 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0:t1_n80, $hdt$_1:t1_n80] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger((_col1 + _col3)) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 2 -2 2 -3 2 -7 2 -8 4 -PREHOOK: query: EXPLAIN EXTENDED -SELECT * FROM -(SELECT key, count(1) FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED -SELECT * FROM -(SELECT key, count(1) FROM T1_n80 GROUP BY key) subq1 -JOIN -(SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n80` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t0` -INNER JOIN (SELECT `key`, `val`, COUNT(*) AS `$f2` -FROM `default`.`t1_n80` -WHERE `key` IS NOT NULL -GROUP BY `key`, `val`) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_1:t1_n80] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n80 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 0 - value expressions: _col1 (type: bigint) - auto parallelism: false - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: 1 - value expressions: _col1 (type: string), _col2 (type: bigint) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: t1_n80 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n80 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n80 - name: default.t1_n80 - Truncated Path -> Alias: - /t1_n80 [$hdt$_0:t1_n80] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types string:bigint:string:string:bigint - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: CREATE TABLE T2_n49(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n49 -POSTHOOK: query: CREATE TABLE T2_n49(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n49 -PREHOOK: query: INSERT OVERWRITE TABLE T2_n49 select key, val from T1_n80 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n80 -PREHOOK: Output: default@t2_n49 -POSTHOOK: query: INSERT OVERWRITE TABLE T2_n49 select key, val from T1_n80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n80 -POSTHOOK: Output: default@t2_n49 -POSTHOOK: Lineage: t2_n49.key SIMPLE [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n49.val SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T2_n49 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T2_n49 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl1_n18 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t2_n49` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n49 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n49 - name: default.t2_n49 - Truncated Path -> Alias: - /t2_n49 [t2_n49] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n18 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T2_n49 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 -SELECT key, count(1) FROM T2_n49 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl1_n18 -POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n18 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n49` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n49 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n49 - name: default.t2_n49 - Truncated Path -> Alias: - /t2_n49 [t2_n49] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: CREATE TABLE outputTbl5_n2(key1 int, key2 int, key3 string, key4 int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl5_n2 -POSTHOOK: query: CREATE TABLE outputTbl5_n2(key1 int, key2 int, key3 string, key4 int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl5_n2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl5_n2 -SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl5_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl5_n2 -SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl5_n2 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, 2 AS `_o__c3`, COUNT(*) AS `_o__c4` -FROM `default`.`t2_n49` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), 2 (type: int), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) - outputColumnNames: key1, key2, key3, key4, cnt - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(key4, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n49 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n49 - name: default.t2_n49 - Truncated Path -> Alias: - /t2_n49 [t2_n49] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct:struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, key4, cnt - Column Types: int, int, string, int, int - Table: default.outputtbl5_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - name: default.outputtbl5_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n2 - name: default.outputtbl5_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n2 -SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl5_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n2 -SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl5_n2 -POSTHOOK: Lineage: outputtbl5_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: outputtbl5_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl5_n2.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl5_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl5_n2.key4 SIMPLE [] -PREHOOK: query: SELECT * FROM outputTbl5_n2 -ORDER BY key1, key2, key3, key4 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl5_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl5_n2 -ORDER BY key1, key2, key3, key4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl5_n2 -#### A masked pattern was here #### -1 1 11 2 1 -2 1 12 2 1 -3 1 13 2 1 -7 1 17 2 1 -8 1 18 2 1 -8 1 28 2 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n49)subq -group by key, constant, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n49)subq -group by key, constant, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -OPTIMIZED SQL: SELECT `key`, 1 AS `constant`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n49` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n49 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n49 - name: default.t2_n49 - Truncated Path -> Alias: - /t2_n49 [t2_n49] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n49)subq -group by key, constant, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n49)subq -group by key, constant, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n49)subq -)subq2 -group by key, constant3, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n2 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n49)subq -)subq2 -group by key, constant3, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -OPTIMIZED SQL: SELECT `key`, 2 AS `constant3`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n49` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 2 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n49 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n49 - name: default.t2_n49 - Truncated Path -> Alias: - /t2_n49 [t2_n49] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n2 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n2 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n2 - name: default.outputtbl4_n2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n49)subq -)subq2 -group by key, constant3, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@outputtbl4_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n49)subq -)subq2 -group by key, constant3, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@outputtbl4_n2 -POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n2 -#### A masked pattern was here #### -1 2 11 1 -2 2 12 1 -3 2 13 1 -7 2 17 1 -8 2 18 1 -8 2 28 1 -PREHOOK: query: CREATE TABLE DEST1_n80(key INT, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST1_n80 -POSTHOOK: query: CREATE TABLE DEST1_n80(key INT, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST1_n80 -PREHOOK: query: CREATE TABLE DEST2_n18(key INT, val STRING, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST2_n18 -POSTHOOK: query: CREATE TABLE DEST2_n18(key INT, val STRING, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST2_n18 -PREHOOK: query: EXPLAIN -FROM T2_n49 -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@dest1_n80 -PREHOOK: Output: default@dest2_n18 -POSTHOOK: query: EXPLAIN -FROM T2_n49 -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@dest1_n80 -POSTHOOK: Output: default@dest2_n18 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n18 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n80 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n80 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest1_n80 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: int, string, int - Table: default.dest2_n18 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n18 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM T2_n49 -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@dest1_n80 -PREHOOK: Output: default@dest2_n18 -POSTHOOK: query: FROM T2_n49 -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@dest1_n80 -POSTHOOK: Output: default@dest2_n18 -POSTHOOK: Lineage: dest1_n80.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: dest1_n80.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: dest2_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n18.val SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: select * from DEST1_n80 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n80 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST1_n80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n80 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: select * from DEST2_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n18 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST2_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n18 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 -PREHOOK: query: EXPLAIN -FROM (select key, val from T2_n49 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@dest1_n80 -PREHOOK: Output: default@dest2_n18 -POSTHOOK: query: EXPLAIN -FROM (select key, val from T2_n49 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@dest1_n80 -POSTHOOK: Output: default@dest2_n18 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-6 - Stage-4 depends on stages: Stage-2 - Stage-5 depends on stages: Stage-1, Stage-4, Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n49 - filterExpr: (UDFToDouble(key) = 8.0D) (type: boolean) - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) = 8.0D) (type: boolean) - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - bucketGroup: true - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Group By Operator - aggregations: count(1) - keys: _col0 (type: string), _col1 (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 537 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n18 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n80 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n80 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest1_n80 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: int, string, int - Table: default.dest2_n18 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n18 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM (select key, val from T2_n49 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n49 -PREHOOK: Output: default@dest1_n80 -PREHOOK: Output: default@dest2_n18 -POSTHOOK: query: FROM (select key, val from T2_n49 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n49 -POSTHOOK: Output: default@dest1_n80 -POSTHOOK: Output: default@dest2_n18 -POSTHOOK: Lineage: dest1_n80.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: dest1_n80.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] -POSTHOOK: Lineage: dest2_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n18.val SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: select * from DEST1_n80 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n80 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST1_n80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n80 -#### A masked pattern was here #### -8 2 -PREHOOK: query: select * from DEST2_n18 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n18 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST2_n18 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n18 -#### A masked pattern was here #### -8 18 1 -8 28 1 diff --git ql/src/test/results/clientpositive/groupby_sort_2.q.out ql/src/test/results/clientpositive/groupby_sort_2.q.out deleted file mode 100644 index a6b2403f47..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_2.q.out +++ /dev/null @@ -1,185 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n51(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (val) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n51 -POSTHOOK: query: CREATE TABLE T1_n51(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (val) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n51 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n51 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n51 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n51 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n51 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n51 select key, val from T1_n51 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n51 -PREHOOK: Output: default@t1_n51 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n51 select key, val from T1_n51 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n51 -POSTHOOK: Output: default@t1_n51 -POSTHOOK: Lineage: t1_n51.key SIMPLE [(t1_n51)t1_n51.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n51.val SIMPLE [(t1_n51)t1_n51.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n10(val string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n10 -POSTHOOK: query: CREATE TABLE outputTbl1_n10(val string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n10 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n10 -SELECT val, count(1) FROM T1_n51 GROUP BY val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n51 -PREHOOK: Output: default@outputtbl1_n10 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n10 -SELECT val, count(1) FROM T1_n51 GROUP BY val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n51 -POSTHOOK: Output: default@outputtbl1_n10 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n51 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val (type: string) - outputColumnNames: val - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - bucketGroup: true - keys: val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n10 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: val, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n10 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: val, cnt - Column Types: string, int - Table: default.outputtbl1_n10 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n10 -SELECT val, count(1) FROM T1_n51 GROUP BY val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n51 -PREHOOK: Output: default@outputtbl1_n10 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n10 -SELECT val, count(1) FROM T1_n51 GROUP BY val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n51 -POSTHOOK: Output: default@outputtbl1_n10 -POSTHOOK: Lineage: outputtbl1_n10.cnt EXPRESSION [(t1_n51)t1_n51.null, ] -POSTHOOK: Lineage: outputtbl1_n10.val SIMPLE [(t1_n51)t1_n51.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n10 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n10 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n10 -#### A masked pattern was here #### -11 1 -12 1 -13 1 -17 1 -18 1 -28 1 diff --git ql/src/test/results/clientpositive/groupby_sort_3.q.out ql/src/test/results/clientpositive/groupby_sort_3.q.out deleted file mode 100644 index e657a28396..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_3.q.out +++ /dev/null @@ -1,370 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n89(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n89 -POSTHOOK: query: CREATE TABLE T1_n89(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n89 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n89 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n89 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n89 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n89 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n89 select key, val from T1_n89 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n89 -PREHOOK: Output: default@t1_n89 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n89 select key, val from T1_n89 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n89 -POSTHOOK: Output: default@t1_n89 -POSTHOOK: Lineage: t1_n89.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n89.val SIMPLE [(t1_n89)t1_n89.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n20(key string, val string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n20 -POSTHOOK: query: CREATE TABLE outputTbl1_n20(key string, val string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n20 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n20 -SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n89 -PREHOOK: Output: default@outputtbl1_n20 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n20 -SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n89 -POSTHOOK: Output: default@outputtbl1_n20 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n89 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n20 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n20 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: string, string, int - Table: default.outputtbl1_n20 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n20 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n20 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n20 -SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n89 -PREHOOK: Output: default@outputtbl1_n20 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n20 -SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n89 -POSTHOOK: Output: default@outputtbl1_n20 -POSTHOOK: Lineage: outputtbl1_n20.cnt EXPRESSION [(t1_n89)t1_n89.null, ] -POSTHOOK: Lineage: outputtbl1_n20.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl1_n20.val SIMPLE [(t1_n89)t1_n89.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n20 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n20 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n20 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 -PREHOOK: query: CREATE TABLE outputTbl2_n7(key string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl2_n7 -POSTHOOK: query: CREATE TABLE outputTbl2_n7(key string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl2_n7 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl2_n7 -SELECT key, count(1) FROM T1_n89 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n89 -PREHOOK: Output: default@outputtbl2_n7 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl2_n7 -SELECT key, count(1) FROM T1_n89 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n89 -POSTHOOK: Output: default@outputtbl2_n7 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n89 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n7 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n7 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: string, int - Table: default.outputtbl2_n7 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n7 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n7 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n7 -SELECT key, count(1) FROM T1_n89 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n89 -PREHOOK: Output: default@outputtbl2_n7 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n7 -SELECT key, count(1) FROM T1_n89 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n89 -POSTHOOK: Output: default@outputtbl2_n7 -POSTHOOK: Lineage: outputtbl2_n7.cnt EXPRESSION [(t1_n89)t1_n89.null, ] -POSTHOOK: Lineage: outputtbl2_n7.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl2_n7 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl2_n7 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl2_n7 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl2_n7 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 diff --git ql/src/test/results/clientpositive/groupby_sort_4.q.out ql/src/test/results/clientpositive/groupby_sort_4.q.out deleted file mode 100644 index cadc717f68..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_4.q.out +++ /dev/null @@ -1,341 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n133(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n133 -POSTHOOK: query: CREATE TABLE T1_n133(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n133 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n133 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n133 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n133 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n133 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n133 select key, val from T1_n133 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n133 -PREHOOK: Output: default@t1_n133 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n133 select key, val from T1_n133 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n133 -POSTHOOK: Output: default@t1_n133 -POSTHOOK: Lineage: t1_n133.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n133.val SIMPLE [(t1_n133)t1_n133.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n31(key STRING, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n31 -POSTHOOK: query: CREATE TABLE outputTbl1_n31(key STRING, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n31 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n31 -SELECT key, count(1) FROM T1_n133 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n133 -PREHOOK: Output: default@outputtbl1_n31 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n31 -SELECT key, count(1) FROM T1_n133 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n133 -POSTHOOK: Output: default@outputtbl1_n31 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n133 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n31 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n31 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: string, int - Table: default.outputtbl1_n31 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n31 -SELECT key, count(1) FROM T1_n133 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n133 -PREHOOK: Output: default@outputtbl1_n31 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n31 -SELECT key, count(1) FROM T1_n133 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n133 -POSTHOOK: Output: default@outputtbl1_n31 -POSTHOOK: Lineage: outputtbl1_n31.cnt EXPRESSION [(t1_n133)t1_n133.null, ] -POSTHOOK: Lineage: outputtbl1_n31.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n31 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n31 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n31 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n31 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: CREATE TABLE outputTbl2_n8(key STRING, val STRING, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl2_n8 -POSTHOOK: query: CREATE TABLE outputTbl2_n8(key STRING, val STRING, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl2_n8 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl2_n8 -SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n133 -PREHOOK: Output: default@outputtbl2_n8 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl2_n8 -SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n133 -POSTHOOK: Output: default@outputtbl2_n8 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n133 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n8 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n8 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: string, string, int - Table: default.outputtbl2_n8 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n8 -SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n133 -PREHOOK: Output: default@outputtbl2_n8 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n8 -SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n133 -POSTHOOK: Output: default@outputtbl2_n8 -POSTHOOK: Lineage: outputtbl2_n8.cnt EXPRESSION [(t1_n133)t1_n133.null, ] -POSTHOOK: Lineage: outputtbl2_n8.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl2_n8.val SIMPLE [(t1_n133)t1_n133.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl2_n8 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl2_n8 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl2_n8 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl2_n8 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 diff --git ql/src/test/results/clientpositive/groupby_sort_6.q.out ql/src/test/results/clientpositive/groupby_sort_6.q.out deleted file mode 100644 index 6bf1057799..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_6.q.out +++ /dev/null @@ -1,883 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n61(key STRING, val STRING) PARTITIONED BY (ds string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n61 -POSTHOOK: query: CREATE TABLE T1_n61(key STRING, val STRING) PARTITIONED BY (ds string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n61 -PREHOOK: query: CREATE TABLE outputTbl1_n15(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n15 -POSTHOOK: query: CREATE TABLE outputTbl1_n15(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n15 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Output: default@outputtbl1_n15 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n61` -WHERE `ds` = '1' -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n61 - filterExpr: (ds = '1') (type: boolean) - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (ds = '1') (type: boolean) - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n15 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Output: default@outputtbl1_n15 -POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] -POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n15 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n15 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n61 PARTITION (ds='2') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n61 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n61 PARTITION (ds='2') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n61 -POSTHOOK: Output: default@t1_n61@ds=2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Output: default@outputtbl1_n15 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n61` -WHERE `ds` = '1' -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n61 - filterExpr: (ds = '1') (type: boolean) - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (ds = '1') (type: boolean) - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n15 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Output: default@outputtbl1_n15 -POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] -POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n15 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n15 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Input: default@t1_n61@ds=2 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Input: default@t1_n61@ds=2 -POSTHOOK: Output: default@outputtbl1_n15 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n61` -WHERE `ds` = '2' -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n61 - filterExpr: (ds = '2') (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: ds=2 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2 - properties: - bucket_count -1 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n61 - numFiles 1 - numRows 0 - partition_columns ds - partition_columns.types string - rawDataSize 0 - serialization.ddl struct t1_n61 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n61 - partition_columns ds - partition_columns.types string - serialization.ddl struct t1_n61 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n61 - name: default.t1_n61 - Truncated Path -> Alias: - /t1_n61/ds=2 [t1_n61] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n15 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n15 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n15 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: PARTIAL Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n61 -PREHOOK: Input: default@t1_n61@ds=2 -PREHOOK: Output: default@outputtbl1_n15 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 -SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n61 -POSTHOOK: Input: default@t1_n61@ds=2 -POSTHOOK: Output: default@outputtbl1_n15 -POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] -POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n15 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n15 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n15 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 diff --git ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out deleted file mode 100644 index 0588f5fbe0..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_skew_1_23.q.out +++ /dev/null @@ -1,8374 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n56(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n56 -POSTHOOK: query: CREATE TABLE T1_n56(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n56 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n56 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n56 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n56 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n56 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n56 select key, val from T1_n56 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@t1_n56 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n56 select key, val from T1_n56 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@t1_n56 -POSTHOOK: Lineage: t1_n56.key SIMPLE [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n56.val SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n13(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n13 -POSTHOOK: query: CREATE TABLE outputTbl1_n13(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n13 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T1_n56 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T1_n56 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T1_n56 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T1_n56 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: CREATE TABLE outputTbl2_n3(key1 int, key2 string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl2_n3 -POSTHOOK: query: CREATE TABLE outputTbl2_n3(key1 int, key2 string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl2_n3 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl2_n3 -SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl2_n3 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl2_n3 -SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl2_n3 -OPTIMIZED SQL: SELECT `key`, `val`, COUNT(*) AS `$f2` -FROM `default`.`t1_n56` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:string:int -#### A masked pattern was here #### - name default.outputtbl2_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n3 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:string:int -#### A masked pattern was here #### - name default.outputtbl2_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n3 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, string, int - Table: default.outputtbl2_n3 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n3 -SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl2_n3 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n3 -SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl2_n3 -POSTHOOK: Lineage: outputtbl2_n3.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl2_n3.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl2_n3.key2 SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl2_n3 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl2_n3 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl2_n3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl2_n3 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: CREATE TABLE outputTbl3_n1(key1 int, key2 int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl3_n1 -POSTHOOK: query: CREATE TABLE outputTbl3_n1(key1 int, key2 int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl3_n1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl3_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl3_n1 -OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, COUNT(*) AS `_o__c2` -FROM `default`.`t1_n56` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: 1 (type: int), UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, int, int - Table: default.outputtbl3_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - name: default.outputtbl3_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - name: default.outputtbl3_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl3_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl3_n1 -POSTHOOK: Lineage: outputtbl3_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl3_n1.key1 SIMPLE [] -POSTHOOK: Lineage: outputtbl3_n1.key2 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl3_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl3_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl3_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl3_n1 -#### A masked pattern was here #### -1 1 1 -1 2 1 -1 3 1 -1 7 1 -1 8 2 -PREHOOK: query: CREATE TABLE outputTbl4_n1(key1 int, key2 int, key3 string, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl4_n1 -POSTHOOK: query: CREATE TABLE outputTbl4_n1(key1 int, key2 int, key3 string, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl4_n1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl4_n1 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t1_n56` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n1 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: final - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl4_n1 -POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl3_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl3_n1 -OPTIMIZED SQL: SELECT `key` AS `$f0`, CAST(`key` AS DOUBLE) + 1 AS `$f1`, COUNT(*) AS `$f2` -FROM `default`.`t1_n56` -GROUP BY `key`, CAST(`key` AS DOUBLE) + 1 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: string), _col1 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: double) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0:t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: double) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: double) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: double) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: key1, key2, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,cnt - columns.comments - columns.types int:int:int -#### A masked pattern was here #### - name default.outputtbl3_n1 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3_n1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, cnt - Column Types: int, int, int - Table: default.outputtbl3_n1 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl3_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 -SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl3_n1 -POSTHOOK: Lineage: outputtbl3_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl3_n1.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl3_n1.key2 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl3_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl3_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl3_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl3_n1 -#### A masked pattern was here #### -1 2 1 -2 3 1 -3 4 1 -7 8 1 -8 9 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT cast(key + key as string), sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -group by key + key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT cast(key + key as string), sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -group by key + key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `_o__c0`, SUM(COUNT(*)) AS `_o__c1` -FROM `default`.`t1_n56` -GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0:t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: double) - mode: partials - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: double) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(CAST( _col0 AS STRING)) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT cast(key + key as string), sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -group by key + key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT cast(key + key as string), sum(cnt) from -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -group by key + key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -14 1 -16 2 -2 1 -4 1 -6 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n56 GROUP BY key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n56 GROUP BY key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -UNION ALL -SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0-subquery1:t1_n56, $hdt$_0-subquery2:t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n56 GROUP BY key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT key, count(1) FROM T1_n56 GROUP BY key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -1 1 -2 1 -2 1 -3 1 -3 1 -7 1 -7 1 -8 2 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) FROM T1_n56 GROUP BY key + key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) FROM T1_n56 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) FROM T1_n56 GROUP BY key + key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -GROUP BY `key` -UNION ALL -SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `key`, COUNT(*) AS `_o__c1` -FROM `default`.`t1_n56` -GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-10 depends on stages: Stage-9 - Stage-2 depends on stages: Stage-10 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 - Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: double) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0-subquery2:$hdt$_0:t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: double) - mode: partials - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: double) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types double,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: double) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: CAST( _col0 AS STRING) (type: string), _col1 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - TableScan - GatherStats: false - Union - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0-subquery1:t1_n56] -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - name: default.outputtbl1_n13 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n56 GROUP BY key + key -) subq1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT * FROM ( -SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key - UNION ALL -SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n56 GROUP BY key + key -) subq1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -14 1 -16 2 -2 1 -2 1 -3 1 -4 1 -6 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`$f1` + `t2`.`$f1` AS `_o__c1` -FROM (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t0` -INNER JOIN (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 0 - value expressions: _col1 (type: bigint) - auto parallelism: false - TableScan - alias: t1_n56 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0:t1_n56, $hdt$_1:t1_n56] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger((_col1 + _col3)) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT subq1.key, subq1.cnt+subq2.cnt FROM -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 2 -2 2 -3 2 -7 2 -8 4 -PREHOOK: query: EXPLAIN EXTENDED -SELECT * FROM -(SELECT key, count(1) FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val) subq2 -ON subq1.key = subq2.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED -SELECT * FROM -(SELECT key, count(1) FROM T1_n56 GROUP BY key) subq1 -JOIN -(SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val) subq2 -ON subq1.key = subq2.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t1_n56` -WHERE `key` IS NOT NULL -GROUP BY `key`) AS `t0` -INNER JOIN (SELECT `key`, `val`, COUNT(*) AS `$f2` -FROM `default`.`t1_n56` -WHERE `key` IS NOT NULL -GROUP BY `key`, `val`) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_1:t1_n56] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: partials - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n56 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - tag: 0 - value expressions: _col1 (type: bigint) - auto parallelism: false - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - tag: 1 - value expressions: _col1 (type: string), _col2 (type: bigint) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10005 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types string,string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: t1_n56 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t1_n56 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t1_n56 - name: default.t1_n56 - Truncated Path -> Alias: - /t1_n56 [$hdt$_0:t1_n56] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types string:bigint:string:string:bigint - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: CREATE TABLE T2_n34(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n34 -POSTHOOK: query: CREATE TABLE T2_n34(key STRING, val STRING) -CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n34 -PREHOOK: query: INSERT OVERWRITE TABLE T2_n34 select key, val from T1_n56 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n56 -PREHOOK: Output: default@t2_n34 -POSTHOOK: query: INSERT OVERWRITE TABLE T2_n34 select key, val from T1_n56 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n56 -POSTHOOK: Output: default@t2_n34 -POSTHOOK: Lineage: t2_n34.key SIMPLE [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n34.val SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T2_n34 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T2_n34 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl1_n13 -OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`t2_n34` -GROUP BY `key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n34 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n34 - name: default.t2_n34 - Truncated Path -> Alias: - /t2_n34 [t2_n34] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: partials - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,cnt - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n13 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n13 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T2_n34 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl1_n13 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 -SELECT key, count(1) FROM T2_n34 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl1_n13 -POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl1_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl1_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl1_n13 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n34` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n34 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n34 - name: default.t2_n34 - Truncated Path -> Alias: - /t2_n34 [t2_n34] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: final - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: CREATE TABLE outputTbl5_n1(key1 int, key2 int, key3 string, key4 int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl5_n1 -POSTHOOK: query: CREATE TABLE outputTbl5_n1(key1 int, key2 int, key3 string, key4 int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl5_n1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl5_n1 -SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl5_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl5_n1 -SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl5_n1 -OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, 2 AS `_o__c3`, COUNT(*) AS `_o__c4` -FROM `default`.`t2_n34` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), 2 (type: int), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) - outputColumnNames: key1, key2, key3, key4, cnt - Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(key4, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n34 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n34 - name: default.t2_n34 - Truncated Path -> Alias: - /t2_n34 [t2_n34] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) - mode: final - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4 - columns.types struct:struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, key4, cnt - Column Types: int, int, string, int, int - Table: default.outputtbl5_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - name: default.outputtbl5_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,key4,cnt - columns.comments - columns.types int:int:string:int:int -#### A masked pattern was here #### - name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl5_n1 - name: default.outputtbl5_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n1 -SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl5_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n1 -SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl5_n1 -POSTHOOK: Lineage: outputtbl5_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: outputtbl5_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl5_n1.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl5_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl5_n1.key4 SIMPLE [] -PREHOOK: query: SELECT * FROM outputTbl5_n1 -ORDER BY key1, key2, key3, key4 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl5_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl5_n1 -ORDER BY key1, key2, key3, key4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl5_n1 -#### A masked pattern was here #### -1 1 11 2 1 -2 1 12 2 1 -3 1 13 2 1 -7 1 17 2 1 -8 1 18 2 1 -8 1 28 2 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n34)subq -group by key, constant, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n34)subq -group by key, constant, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -OPTIMIZED SQL: SELECT `key`, 1 AS `constant`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n34` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n34 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n34 - name: default.t2_n34 - Truncated Path -> Alias: - /t2_n34 [t2_n34] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: final - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n34)subq -group by key, constant, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -SELECT key, constant, val, count(1) from -(SELECT key, 1 as constant, val from T2_n34)subq -group by key, constant, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -1 1 11 1 -2 1 12 1 -3 1 13 1 -7 1 17 1 -8 1 18 1 -8 1 28 1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n34)subq -)subq2 -group by key, constant3, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE outputTbl4_n1 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n34)subq -)subq2 -group by key, constant3, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -OPTIMIZED SQL: SELECT `key`, 2 AS `constant3`, `val`, COUNT(*) AS `_o__c3` -FROM `default`.`t2_n34` -GROUP BY `key`, `val` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), 2 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) - outputColumnNames: key1, key2, key3, cnt - Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: t2_n34 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} - SORTBUCKETCOLSPREFIX TRUE - bucket_count 2 - bucket_field_name key,val - bucketing_version 2 - column.name.delimiter , - columns key,val - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n34 - name: default.t2_n34 - Truncated Path -> Alias: - /t2_n34 [t2_n34] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: final - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key1, key2, key3, cnt - Column Types: int, int, string, int - Table: default.outputtbl4_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key1,key2,key3,cnt - columns.comments - columns.types int:int:string:int -#### A masked pattern was here #### - name default.outputtbl4_n1 - numFiles 1 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl4_n1 - name: default.outputtbl4_n1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n34)subq -)subq2 -group by key, constant3, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@outputtbl4_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 -select key, constant3, val, count(1) from -( -SELECT key, constant as constant2, val, 2 as constant3 from -(SELECT key, 1 as constant, val from T2_n34)subq -)subq2 -group by key, constant3, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@outputtbl4_n1 -POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] -POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM outputTbl4_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM outputTbl4_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@outputtbl4_n1 -#### A masked pattern was here #### -1 2 11 1 -2 2 12 1 -3 2 13 1 -7 2 17 1 -8 2 18 1 -8 2 28 1 -PREHOOK: query: CREATE TABLE DEST1_n57(key INT, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST1_n57 -POSTHOOK: query: CREATE TABLE DEST1_n57(key INT, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST1_n57 -PREHOOK: query: CREATE TABLE DEST2_n12(key INT, val STRING, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@DEST2_n12 -POSTHOOK: query: CREATE TABLE DEST2_n12(key INT, val STRING, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@DEST2_n12 -PREHOOK: query: EXPLAIN -FROM T2_n34 -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@dest1_n57 -PREHOOK: Output: default@dest2_n12 -POSTHOOK: query: EXPLAIN -FROM T2_n34 -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@dest1_n57 -POSTHOOK: Output: default@dest2_n12 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0, Stage-5, Stage-7 - Stage-5 depends on stages: Stage-3 - Stage-6 depends on stages: Stage-1, Stage-5, Stage-7 - Stage-1 depends on stages: Stage-2 - Stage-7 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n12 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: partials - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n57 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n57 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest1_n57 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: int, string, int - Table: default.dest2_n12 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n12 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM T2_n34 -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@dest1_n57 -PREHOOK: Output: default@dest2_n12 -POSTHOOK: query: FROM T2_n34 -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@dest1_n57 -POSTHOOK: Output: default@dest2_n12 -POSTHOOK: Lineage: dest1_n57.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: dest1_n57.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n12.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: dest2_n12.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n12.val SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: select * from DEST1_n57 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n57 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST1_n57 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n57 -#### A masked pattern was here #### -1 1 -2 1 -3 1 -7 1 -8 2 -PREHOOK: query: select * from DEST2_n12 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n12 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST2_n12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n12 -#### A masked pattern was here #### -1 11 1 -2 12 1 -3 13 1 -7 17 1 -8 18 1 -8 28 1 -PREHOOK: query: EXPLAIN -FROM (select key, val from T2_n34 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@dest1_n57 -PREHOOK: Output: default@dest2_n12 -POSTHOOK: query: EXPLAIN -FROM (select key, val from T2_n34 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@dest1_n57 -POSTHOOK: Output: default@dest2_n12 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0, Stage-5, Stage-7 - Stage-5 depends on stages: Stage-3 - Stage-6 depends on stages: Stage-1, Stage-5, Stage-7 - Stage-1 depends on stages: Stage-2 - Stage-7 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: t2_n34 - filterExpr: (UDFToDouble(key) = 8.0D) (type: boolean) - Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) = 8.0D) (type: boolean) - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(1) - bucketGroup: true - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: rand() (type: double) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Group By Operator - aggregations: count(1) - keys: _col0 (type: string), _col1 (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 537 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n12 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: partials - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n57 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n57 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest1_n57 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val, cnt - Column Types: int, string, int - Table: default.dest2_n12 - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n12 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: true - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM (select key, val from T2_n34 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n34 -PREHOOK: Output: default@dest1_n57 -PREHOOK: Output: default@dest2_n12 -POSTHOOK: query: FROM (select key, val from T2_n34 where key = 8) x -INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key -INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n34 -POSTHOOK: Output: default@dest1_n57 -POSTHOOK: Output: default@dest2_n12 -POSTHOOK: Lineage: dest1_n57.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: dest1_n57.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n12.cnt EXPRESSION [(t2_n34)t2_n34.null, ] -POSTHOOK: Lineage: dest2_n12.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n12.val SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: select * from DEST1_n57 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n57 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST1_n57 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n57 -#### A masked pattern was here #### -8 2 -PREHOOK: query: select * from DEST2_n12 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n12 -#### A masked pattern was here #### -POSTHOOK: query: select * from DEST2_n12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n12 -#### A masked pattern was here #### -8 18 1 -8 28 1 diff --git ql/src/test/results/clientpositive/groupby_sort_test_1.q.out ql/src/test/results/clientpositive/groupby_sort_test_1.q.out deleted file mode 100644 index 405374af77..0000000000 --- ql/src/test/results/clientpositive/groupby_sort_test_1.q.out +++ /dev/null @@ -1,173 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n164(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n164 -POSTHOOK: query: CREATE TABLE T1_n164(key STRING, val STRING) -CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n164 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n164 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n164 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n164 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n164 -PREHOOK: query: INSERT OVERWRITE TABLE T1_n164 select key, val from T1_n164 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n164 -PREHOOK: Output: default@t1_n164 -POSTHOOK: query: INSERT OVERWRITE TABLE T1_n164 select key, val from T1_n164 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n164 -POSTHOOK: Output: default@t1_n164 -POSTHOOK: Lineage: t1_n164.key SIMPLE [(t1_n164)t1_n164.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t1_n164.val SIMPLE [(t1_n164)t1_n164.FieldSchema(name:val, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE outputTbl1_n35(key int, cnt int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@outputTbl1_n35 -POSTHOOK: query: CREATE TABLE outputTbl1_n35(key int, cnt int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@outputTbl1_n35 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n35 -SELECT key, count(1) FROM T1_n164 GROUP BY key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n164 -PREHOOK: Output: default@outputtbl1_n35 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE outputTbl1_n35 -SELECT key, count(1) FROM T1_n164 GROUP BY key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n164 -POSTHOOK: Output: default@outputtbl1_n35 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n164 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n35 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n35 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.outputtbl1_n35 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n35 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n35 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - diff --git ql/src/test/results/clientpositive/hashjoin.q.out ql/src/test/results/clientpositive/hashjoin.q.out deleted file mode 100644 index 27194bf841..0000000000 --- ql/src/test/results/clientpositive/hashjoin.q.out +++ /dev/null @@ -1,566 +0,0 @@ -PREHOOK: query: CREATE TABLE t_hashjoin_big( - cint int, - cvarchar varchar(50), - cdouble double, - a int -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t_hashjoin_big -POSTHOOK: query: CREATE TABLE t_hashjoin_big( - cint int, - cvarchar varchar(50), - cdouble double, - a int -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t_hashjoin_big -PREHOOK: query: CREATE TABLE t_hashjoin_small( - cint int, - cvarchar varchar(50), - cdouble double -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@t_hashjoin_small -POSTHOOK: query: CREATE TABLE t_hashjoin_small( - cint int, - cvarchar varchar(50), - cdouble double -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@t_hashjoin_small -PREHOOK: query: INSERT INTO t_hashjoin_big VALUES -(5, 'two', 3.0, 1), -(6, 'two', 1.5, 2), -(NULL, NULL, NULL, NULL), -(7, 'eight', 4.2, 3), (7, 'eight', 4.2, 4), (7, 'eight', 4.2, 5), -(5, 'one', 2.8, 6), (5, 'one', 2.8, 7), (5, 'one', 2.8, 8) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@t_hashjoin_big -POSTHOOK: query: INSERT INTO t_hashjoin_big VALUES -(5, 'two', 3.0, 1), -(6, 'two', 1.5, 2), -(NULL, NULL, NULL, NULL), -(7, 'eight', 4.2, 3), (7, 'eight', 4.2, 4), (7, 'eight', 4.2, 5), -(5, 'one', 2.8, 6), (5, 'one', 2.8, 7), (5, 'one', 2.8, 8) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@t_hashjoin_big -POSTHOOK: Lineage: t_hashjoin_big.a SCRIPT [] -POSTHOOK: Lineage: t_hashjoin_big.cdouble SCRIPT [] -POSTHOOK: Lineage: t_hashjoin_big.cint SCRIPT [] -POSTHOOK: Lineage: t_hashjoin_big.cvarchar SCRIPT [] -PREHOOK: query: INSERT INTO t_hashjoin_small VALUES -(7, 'two', 1.5), -(5, 'two', 4.2), -(NULL, NULL, NULL), -(5, 'one', 1.1), (5, 'one', 1.1) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@t_hashjoin_small -POSTHOOK: query: INSERT INTO t_hashjoin_small VALUES -(7, 'two', 1.5), -(5, 'two', 4.2), -(NULL, NULL, NULL), -(5, 'one', 1.1), (5, 'one', 1.1) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@t_hashjoin_small -POSTHOOK: Lineage: t_hashjoin_small.cdouble SCRIPT [] -POSTHOOK: Lineage: t_hashjoin_small.cint SCRIPT [] -POSTHOOK: Lineage: t_hashjoin_small.cvarchar SCRIPT [] -PREHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - z - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - z - TableScan - alias: z - filterExpr: cint is not null (type: boolean) - Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cint is not null (type: boolean) - Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 cint (type: int) - 1 cint (type: int) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: cint is not null (type: boolean) - Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cint is not null (type: boolean) - Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 cint (type: int) - 1 cint (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -5 two 3.0 1 5 two 4.2 -5 two 3.0 1 5 one 1.1 -5 two 3.0 1 5 one 1.1 -7 eight 4.2 3 7 two 1.5 -7 eight 4.2 4 7 two 1.5 -7 eight 4.2 5 7 two 1.5 -5 one 2.8 6 5 two 4.2 -5 one 2.8 6 5 one 1.1 -5 one 2.8 6 5 one 1.1 -5 one 2.8 7 5 two 4.2 -5 one 2.8 7 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 8 5 two 4.2 -5 one 2.8 8 5 one 1.1 -5 one 2.8 8 5 one 1.1 -PREHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - z - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - z - TableScan - alias: z - filterExpr: cvarchar is not null (type: boolean) - Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cvarchar is not null (type: boolean) - Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 cvarchar (type: varchar(50)) - 1 cvarchar (type: varchar(50)) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: cvarchar is not null (type: boolean) - Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cvarchar is not null (type: boolean) - Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 cvarchar (type: varchar(50)) - 1 cvarchar (type: varchar(50)) - outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -5 two 3.0 1 7 two 1.5 -5 two 3.0 1 5 two 4.2 -6 two 1.5 2 7 two 1.5 -6 two 1.5 2 5 two 4.2 -5 one 2.8 6 5 one 1.1 -5 one 2.8 6 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 8 5 one 1.1 -5 one 2.8 8 5 one 1.1 -PREHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - z - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - z - TableScan - alias: z - filterExpr: (cint is not null and cvarchar is not null) (type: boolean) - Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (cint is not null and cvarchar is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 cint (type: int), cvarchar (type: varchar(50)) - 1 cint (type: int), cvarchar (type: varchar(50)) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (cint is not null and cvarchar is not null) (type: boolean) - Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (cint is not null and cvarchar is not null) (type: boolean) - Statistics: Num rows: 7 Data size: 721 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 cint (type: int), cvarchar (type: varchar(50)) - 1 cint (type: int), cvarchar (type: varchar(50)) - outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 - Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -5 two 3.0 1 5 two 4.2 -5 one 2.8 6 5 one 1.1 -5 one 2.8 6 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 8 5 one 1.1 -5 one 2.8 8 5 one 1.1 -PREHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - z - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - z - TableScan - alias: z - Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 cint (type: int), cvarchar (type: varchar(50)) - 1 cint (type: int), cvarchar (type: varchar(50)) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 cint (type: int), cvarchar (type: varchar(50)) - 1 cint (type: int), cvarchar (type: varchar(50)) - outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 - Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -5 two 3.0 1 5 two 4.2 -6 two 1.5 2 NULL NULL NULL -NULL NULL NULL NULL NULL NULL NULL -7 eight 4.2 3 NULL NULL NULL -7 eight 4.2 4 NULL NULL NULL -7 eight 4.2 5 NULL NULL NULL -5 one 2.8 6 5 one 1.1 -5 one 2.8 6 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 7 5 one 1.1 -5 one 2.8 8 5 one 1.1 -5 one 2.8 8 5 one 1.1 -PREHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - z - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - z - TableScan - alias: z - filterExpr: cdouble is not null (type: boolean) - Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cdouble is not null (type: boolean) - Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 cdouble (type: double) - 1 cdouble (type: double) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: cdouble is not null (type: boolean) - Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: cdouble is not null (type: boolean) - Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 cdouble (type: double) - 1 cdouble (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 - Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) -PREHOOK: type: QUERY -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Input: default@t_hashjoin_small -#### A masked pattern was here #### -6 two 1.5 2 7 two 1.5 -7 eight 4.2 3 5 two 4.2 -7 eight 4.2 4 5 two 4.2 -7 eight 4.2 5 5 two 4.2 -PREHOOK: query: DROP TABLE t_hashjoin_big -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@t_hashjoin_big -PREHOOK: Output: default@t_hashjoin_big -POSTHOOK: query: DROP TABLE t_hashjoin_big -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@t_hashjoin_big -POSTHOOK: Output: default@t_hashjoin_big -PREHOOK: query: DROP TABLE t_hashjoin_small -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@t_hashjoin_small -PREHOOK: Output: default@t_hashjoin_small -POSTHOOK: query: DROP TABLE t_hashjoin_small -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@t_hashjoin_small -POSTHOOK: Output: default@t_hashjoin_small diff --git ql/src/test/results/clientpositive/having2.q.out ql/src/test/results/clientpositive/having2.q.out deleted file mode 100644 index 74bb312940..0000000000 --- ql/src/test/results/clientpositive/having2.q.out +++ /dev/null @@ -1,662 +0,0 @@ -PREHOOK: query: CREATE TABLE TestV1_Staples ( - Item_Count INT, - Ship_Priority STRING, - Order_Priority STRING, - Order_Status STRING, - Order_Quantity DOUBLE, - Sales_Total DOUBLE, - Discount DOUBLE, - Tax_Rate DOUBLE, - Ship_Mode STRING, - Fill_Time DOUBLE, - Gross_Profit DOUBLE, - Price DOUBLE, - Ship_Handle_Cost DOUBLE, - Employee_Name STRING, - Employee_Dept STRING, - Manager_Name STRING, - Employee_Yrs_Exp DOUBLE, - Employee_Salary DOUBLE, - Customer_Name STRING, - Customer_State STRING, - Call_Center_Region STRING, - Customer_Balance DOUBLE, - Customer_Segment STRING, - Prod_Type1 STRING, - Prod_Type2 STRING, - Prod_Type3 STRING, - Prod_Type4 STRING, - Product_Name STRING, - Product_Container STRING, - Ship_Promo STRING, - Supplier_Name STRING, - Supplier_Balance DOUBLE, - Supplier_Region STRING, - Supplier_State STRING, - Order_ID STRING, - Order_Year INT, - Order_Month INT, - Order_Day INT, - Order_Date_ STRING, - Order_Quarter STRING, - Product_Base_Margin DOUBLE, - Product_ID STRING, - Receive_Time DOUBLE, - Received_Date_ STRING, - Ship_Date_ STRING, - Ship_Charge DOUBLE, - Total_Cycle_Time DOUBLE, - Product_In_Stock STRING, - PID INT, - Market_Segment STRING - ) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@TestV1_Staples -POSTHOOK: query: CREATE TABLE TestV1_Staples ( - Item_Count INT, - Ship_Priority STRING, - Order_Priority STRING, - Order_Status STRING, - Order_Quantity DOUBLE, - Sales_Total DOUBLE, - Discount DOUBLE, - Tax_Rate DOUBLE, - Ship_Mode STRING, - Fill_Time DOUBLE, - Gross_Profit DOUBLE, - Price DOUBLE, - Ship_Handle_Cost DOUBLE, - Employee_Name STRING, - Employee_Dept STRING, - Manager_Name STRING, - Employee_Yrs_Exp DOUBLE, - Employee_Salary DOUBLE, - Customer_Name STRING, - Customer_State STRING, - Call_Center_Region STRING, - Customer_Balance DOUBLE, - Customer_Segment STRING, - Prod_Type1 STRING, - Prod_Type2 STRING, - Prod_Type3 STRING, - Prod_Type4 STRING, - Product_Name STRING, - Product_Container STRING, - Ship_Promo STRING, - Supplier_Name STRING, - Supplier_Balance DOUBLE, - Supplier_Region STRING, - Supplier_State STRING, - Order_ID STRING, - Order_Year INT, - Order_Month INT, - Order_Day INT, - Order_Date_ STRING, - Order_Quarter STRING, - Product_Base_Margin DOUBLE, - Product_ID STRING, - Receive_Time DOUBLE, - Received_Date_ STRING, - Ship_Date_ STRING, - Ship_Charge DOUBLE, - Total_Cycle_Time DOUBLE, - Product_In_Stock STRING, - PID INT, - Market_Segment STRING - ) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@TestV1_Staples -PREHOOK: query: explain -SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( -(COUNT(s1.discount) <= 822) AND -(SUM(customer_balance) <= 4074689.000000041) -) -PREHOOK: type: QUERY -PREHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( -(COUNT(s1.discount) <= 822) AND -(SUM(customer_balance) <= 4074689.000000041) -) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: order_quantity (type: double), discount (type: double), customer_name (type: string), customer_balance (type: double) - outputColumnNames: order_quantity, discount, customer_name, customer_balance - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(customer_balance), sum(order_quantity), count(discount) - keys: customer_name (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col3 <= 822L) and (_col1 <= 4074689.000000041D)) (type: boolean) - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( -(SUM(customer_balance) <= 4074689.000000041) -AND (COUNT(s1.discount) <= 822) -) -PREHOOK: type: QUERY -PREHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( -(SUM(customer_balance) <= 4074689.000000041) -AND (COUNT(s1.discount) <= 822) -) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: order_quantity (type: double), discount (type: double), customer_name (type: string), customer_balance (type: double) - outputColumnNames: order_quantity, discount, customer_name, customer_balance - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(customer_balance), sum(order_quantity), count(discount) - keys: customer_name (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col1 <= 4074689.000000041D) and (_col3 <= 822L)) (type: boolean) - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -GROUP BY s1.customer_name -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -GROUP BY s1.customer_name -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - filterExpr: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: double), _col2 (type: double) - TableScan - alias: s2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(_col2), count(_col4), sum(_col0), count(_col0) - keys: _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: double), _col2 (type: bigint), _col3 (type: double), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1), sum(VALUE._col2), count(VALUE._col3) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col1 <= 4074689.000000041D) and (_col2 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -GROUP BY s1.customer_name, s1.customer_name -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -GROUP BY s1.customer_name, s1.customer_name -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - filterExpr: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: double), _col2 (type: double) - TableScan - alias: s2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(_col2), sum(_col0), count(_col0), count(_col4) - keys: _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2), count(VALUE._col3) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) - outputColumnNames: _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col2 <= 4074689.000000041D) and (_col5 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -SELECT distinct s1.customer_name as x, s1.customer_name as y -FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT distinct s1.customer_name as x, s1.customer_name as y -FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key -HAVING ( -(SUM(s1.customer_balance) <= 4074689.000000041) -AND (AVG(s1.discount) <= 822) -AND (COUNT(s2.value) > 4) -) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@testv1_staples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - filterExpr: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: customer_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: double), _col2 (type: double) - TableScan - alias: s2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: sum(_col2), sum(_col0), count(_col0), count(_col4) - keys: _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2), count(VALUE._col3) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) - outputColumnNames: _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col2 <= 4074689.000000041D) and (_col5 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out deleted file mode 100644 index 255823c08b..0000000000 --- ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out +++ /dev/null @@ -1,140 +0,0 @@ -PREHOOK: query: create table implicit_cast_during_insert (c1 int, c2 string) - partitioned by (p1 string) stored as orc -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@implicit_cast_during_insert -POSTHOOK: query: create table implicit_cast_during_insert (c1 int, c2 string) - partitioned by (p1 string) stored as orc -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@implicit_cast_during_insert -PREHOOK: query: explain -insert overwrite table implicit_cast_during_insert partition (p1) - select key, value, key key1 from (select * from src where key in (0,1)) q - distribute by key1 sort by key1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@implicit_cast_during_insert -POSTHOOK: query: explain -insert overwrite table implicit_cast_during_insert partition (p1) - select key, value, key key1 from (select * from src where key in (0,1)) q - distribute by key1 sort by key1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key) IN (0, 1) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key) IN (0, 1) (type: boolean) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(KEY.reducesinkkey0) (type: int), VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.implicit_cast_during_insert - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: c1, c2, p1 - Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') - keys: p1 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 1902 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 1902 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1902 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Move Operator - tables: - partition: - p1 - replace: true - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.implicit_cast_during_insert - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2 - Column Types: int, string - Table: default.implicit_cast_during_insert - -PREHOOK: query: insert overwrite table implicit_cast_during_insert partition (p1) - select key, value, key key1 from (select * from src where key in (0,1)) q - distribute by key1 sort by key1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@implicit_cast_during_insert -POSTHOOK: query: insert overwrite table implicit_cast_during_insert partition (p1) - select key, value, key key1 from (select * from src where key in (0,1)) q - distribute by key1 sort by key1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@implicit_cast_during_insert@p1=0 -POSTHOOK: Lineage: implicit_cast_during_insert PARTITION(p1=0).c1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: implicit_cast_during_insert PARTITION(p1=0).c2 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from implicit_cast_during_insert -PREHOOK: type: QUERY -PREHOOK: Input: default@implicit_cast_during_insert -PREHOOK: Input: default@implicit_cast_during_insert@p1=0 -#### A masked pattern was here #### -POSTHOOK: query: select * from implicit_cast_during_insert -POSTHOOK: type: QUERY -POSTHOOK: Input: default@implicit_cast_during_insert -POSTHOOK: Input: default@implicit_cast_during_insert@p1=0 -#### A masked pattern was here #### -0 val_0 0 -0 val_0 0 -0 val_0 0 -PREHOOK: query: drop table implicit_cast_during_insert -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@implicit_cast_during_insert -PREHOOK: Output: default@implicit_cast_during_insert -POSTHOOK: query: drop table implicit_cast_during_insert -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@implicit_cast_during_insert -POSTHOOK: Output: default@implicit_cast_during_insert diff --git ql/src/test/results/clientpositive/in_typecheck_char.q.out ql/src/test/results/clientpositive/in_typecheck_char.q.out deleted file mode 100644 index d2fcdf48f0..0000000000 --- ql/src/test/results/clientpositive/in_typecheck_char.q.out +++ /dev/null @@ -1,416 +0,0 @@ -PREHOOK: query: create table ax(s char(1),t char(10)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@ax -POSTHOOK: query: create table ax(s char(1),t char(10)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@ax -PREHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@ax -POSTHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@ax -POSTHOOK: Lineage: ax.s SCRIPT [] -POSTHOOK: Lineage: ax.t SCRIPT [] -PREHOOK: query: explain -select 'expected 2',count(*) from ax where s = 'a' and t = 'a' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain -select 'expected 2',count(*) from ax where s = 'a' and t = 'a' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: ((s = 'a') and (t = 'a ')) (type: boolean) - Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((s = 'a') and (t = 'a ')) (type: boolean) - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'expected 2' (type: string), _col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 'expected 2',count(*) from ax where s = 'a' and t = 'a' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 2',count(*) from ax where s = 'a' and t = 'a' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 2 2 -PREHOOK: query: explain -select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain -select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) - Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'expected 3' (type: string), _col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 3 3 -PREHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 2 2 -PREHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 2 2 -PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 0 0 -PREHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),(null, 'bb')) -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),(null, 'bb')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 2 2 -PREHOOK: query: explain select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: (((CAST( s AS STRING) = 'a') and (CAST( t AS STRING) = 'a')) or (null and (CAST( t AS STRING) = 'bb'))) is null (type: boolean) - Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((CAST( s AS STRING) = 'a') and (CAST( t AS STRING) = 'a')) or (null and (CAST( t AS STRING) = 'bb'))) is null (type: boolean) - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'expected 1' (type: string), _col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 1 1 -PREHOOK: query: explain cbo select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain cbo select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -CBO PLAN: -HiveAggregate(group=[{}], agg#0=[count()]) - HiveFilter(condition=[IN($1, _UTF-16LE'a ', _UTF-16LE'bb ', _UTF-16LE'aa ', _UTF-16LE'bbb ', _UTF-16LE'ab ', _UTF-16LE'ba ', _UTF-16LE'aaa ', _UTF-16LE'bbb ', _UTF-16LE'abc ', _UTF-16LE'bc ', _UTF-16LE'ac ', _UTF-16LE'bca ', _UTF-16LE'cab ', _UTF-16LE'cb ', _UTF-16LE'ca ', _UTF-16LE'cbc ', _UTF-16LE'z ')]) - HiveTableScan(table=[[default, ax]], table:alias=[ax]) - -PREHOOK: query: explain select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: (t) IN ('a ', 'bb ', 'aa ', 'bbb ', 'ab ', 'ba ', 'aaa ', 'bbb ', 'abc ', 'bc ', 'ac ', 'bca ', 'cab ', 'cb ', 'ca ', 'cbc ', 'z ') (type: boolean) - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (t) IN ('a ', 'bb ', 'aa ', 'bbb ', 'ab ', 'ba ', 'aaa ', 'bbb ', 'abc ', 'bc ', 'ac ', 'bca ', 'cab ', 'cb ', 'ca ', 'cbc ', 'z ') (type: boolean) - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain cbo select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain cbo select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -CBO PLAN: -HiveAggregate(group=[{}], agg#0=[count()]) - HiveFilter(condition=[OR(=($1, _UTF-16LE'a '), =($1, _UTF-16LE'bb '), =($1, _UTF-16LE'aa '), =($1, _UTF-16LE'bbb '), =($1, _UTF-16LE'ab '), =($1, _UTF-16LE'ba '), =($1, _UTF-16LE'aaa '), =($1, _UTF-16LE'abc '), =($1, _UTF-16LE'bc '), =($1, _UTF-16LE'ac '), =($1, _UTF-16LE'bca '), =($1, _UTF-16LE'cab '), =($1, _UTF-16LE'cb '), =($1, _UTF-16LE'ca '), =($1, _UTF-16LE'cbc '), =($1, _UTF-16LE'z '))]) - HiveTableScan(table=[[default, ax]], table:alias=[ax]) - -PREHOOK: query: explain select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain select count(*) from ax where t in -('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: ((t = 'a ') or (t = 'bb ') or (t = 'aa ') or (t = 'bbb ') or (t = 'ab ') or (t = 'ba ') or (t = 'aaa ') or (t = 'abc ') or (t = 'bc ') or (t = 'ac ') or (t = 'bca ') or (t = 'cab ') or (t = 'cb ') or (t = 'ca ') or (t = 'cbc ') or (t = 'z ')) (type: boolean) - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((t = 'a ') or (t = 'bb ') or (t = 'aa ') or (t = 'bbb ') or (t = 'ab ') or (t = 'ba ') or (t = 'aaa ') or (t = 'abc ') or (t = 'bc ') or (t = 'ac ') or (t = 'bca ') or (t = 'cab ') or (t = 'cb ') or (t = 'ca ') or (t = 'cbc ') or (t = 'z ')) (type: boolean) - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/in_typecheck_pointlook.q.out ql/src/test/results/clientpositive/in_typecheck_pointlook.q.out deleted file mode 100644 index 7274a24da2..0000000000 --- ql/src/test/results/clientpositive/in_typecheck_pointlook.q.out +++ /dev/null @@ -1,185 +0,0 @@ -PREHOOK: query: create table customer_demographics (cd_marital_status char(1), cd_education_status char(20)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@customer_demographics -POSTHOOK: query: create table customer_demographics (cd_marital_status char(1), cd_education_status char(20)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@customer_demographics -PREHOOK: query: insert into customer_demographics values -('M','Unknown'), -('W','Advanced Degree'), -('W','Advanced Degree '), -('W',' Advanced Degree') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@customer_demographics -POSTHOOK: query: insert into customer_demographics values -('M','Unknown'), -('W','Advanced Degree'), -('W','Advanced Degree '), -('W',' Advanced Degree') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@customer_demographics -POSTHOOK: Lineage: customer_demographics.cd_education_status SCRIPT [] -POSTHOOK: Lineage: customer_demographics.cd_marital_status SCRIPT [] -PREHOOK: query: explain -select count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -PREHOOK: type: QUERY -PREHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -POSTHOOK: query: explain -select count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: customer_demographics - filterExpr: (((cd_marital_status = 'M') and (cd_education_status = 'Unknown ')) or ((cd_marital_status = 'W') and (cd_education_status = 'Advanced Degree '))) (type: boolean) - Statistics: Num rows: 4 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((cd_marital_status = 'M') and (cd_education_status = 'Unknown ')) or ((cd_marital_status = 'W') and (cd_education_status = 'Advanced Degree '))) (type: boolean) - Statistics: Num rows: 2 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 2 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select '3 is expected:',count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -PREHOOK: type: QUERY -PREHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -POSTHOOK: query: select '3 is expected:',count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -3 is expected: 3 -PREHOOK: query: explain -select count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -PREHOOK: type: QUERY -PREHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -POSTHOOK: query: explain -select count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: customer_demographics - filterExpr: (struct(cd_marital_status,cd_education_status)) IN (const struct('M','Unknown '), const struct('W','Advanced Degree ')) (type: boolean) - Statistics: Num rows: 4 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (struct(cd_marital_status,cd_education_status)) IN (const struct('M','Unknown '), const struct('W','Advanced Degree ')) (type: boolean) - Statistics: Num rows: 1 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select '3 is expected:',count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -PREHOOK: type: QUERY -PREHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -POSTHOOK: query: select '3 is expected:',count(1) -from customer_demographics -where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') -or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@customer_demographics -#### A masked pattern was here #### -3 is expected: 3 diff --git ql/src/test/results/clientpositive/in_typecheck_varchar.q.out ql/src/test/results/clientpositive/in_typecheck_varchar.q.out deleted file mode 100644 index 2c73d020bd..0000000000 --- ql/src/test/results/clientpositive/in_typecheck_varchar.q.out +++ /dev/null @@ -1,189 +0,0 @@ -PREHOOK: query: create table ax(s varchar(1),t varchar(10)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@ax -POSTHOOK: query: create table ax(s varchar(1),t varchar(10)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@ax -PREHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@ax -POSTHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@ax -POSTHOOK: Lineage: ax.s SCRIPT [] -POSTHOOK: Lineage: ax.t SCRIPT [] -PREHOOK: query: explain -select 'expected 1',count(*) from ax where s = 'a' and t = 'a' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain -select 'expected 1',count(*) from ax where s = 'a' and t = 'a' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: ((s = 'a') and (t = 'a')) (type: boolean) - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((s = 'a') and (t = 'a')) (type: boolean) - Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'expected 1' (type: string), _col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 'expected 1',count(*) from ax where s = 'a' and t = 'a' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 1',count(*) from ax where s = 'a' and t = 'a' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 1 1 -PREHOOK: query: explain -select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: explain -select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: ax - filterExpr: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) - Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) - Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'expected 2' (type: string), _col0 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 2 2 -PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 0 0 -PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 0 0 -PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' -PREHOOK: type: QUERY -PREHOOK: Input: default@ax -#### A masked pattern was here #### -POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ax -#### A masked pattern was here #### -expected 0 0 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_num_buckets.q.out ql/src/test/results/clientpositive/infer_bucket_sort_num_buckets.q.out deleted file mode 100644 index cd1996433c..0000000000 --- ql/src/test/results/clientpositive/infer_bucket_sort_num_buckets.q.out +++ /dev/null @@ -1,532 +0,0 @@ -PREHOOK: query: CREATE TABLE test_table_n0 (key INT, value STRING) PARTITIONED BY (ds STRING, hr STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_table_n0 -POSTHOOK: query: CREATE TABLE test_table_n0 (key INT, value STRING) PARTITIONED BY (ds STRING, hr STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@test_table_n0 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) -SELECT key2, value, cast(hr as int) FROM -(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 2) as hr -FROM srcpart -WHERE ds = '2008-04-08') a -DISTRIBUTE BY key2 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@test_table_n0@ds=2008-04-08 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) -SELECT key2, value, cast(hr as int) FROM -(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 2) as hr -FROM srcpart -WHERE ds = '2008-04-08') a -DISTRIBUTE BY key2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-4 depends on stages: Stage-0, Stage-2 - Stage-3 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: if(((key % 3) < 2), 0, 1) (type: int), value (type: string), UDFToInteger((key % 2)) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 99000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1000 Data size: 99000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [a:srcpart] - /srcpart/ds=2008-04-08/hr=12 [a:srcpart] - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), CAST( VALUE._col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, hr - Statistics: Num rows: 1000 Data size: 373000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: '2008-04-08' (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 316 Data size: 360872 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types int,string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: '2008-04-08' (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: '2008-04-08' (type: string), _col1 (type: string) - Statistics: Num rows: 316 Data size: 360872 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: '2008-04-08' (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), '2008-04-08' (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-4 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.test_table_n0 - Is Table Level Stats: false - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col2 (type: string) - null sort order: a - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: int), _col1 (type: string) - auto parallelism: false - Path -> Bucketed Columns: -#### A masked pattern was here #### - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types int,string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types int,string,string - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Path -> Sorted Columns: -#### A masked pattern was here #### - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - Dp Sort State: PARTITION_SORTED - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.test_table_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct test_table_n0 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008-04-08 - hr - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.test_table_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct test_table_n0 { i32 key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_n0 - -PREHOOK: query: INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) -SELECT key2, value, cast(hr as int) FROM -(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 3 % 2) as hr -FROM srcpart -WHERE ds = '2008-04-08') a -DISTRIBUTE BY key2 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@test_table_n0@ds=2008-04-08 -POSTHOOK: query: INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) -SELECT key2, value, cast(hr as int) FROM -(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 3 % 2) as hr -FROM srcpart -WHERE ds = '2008-04-08') a -DISTRIBUTE BY key2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@test_table_n0@ds=2008-04-08/hr=0 -POSTHOOK: Output: default@test_table_n0@ds=2008-04-08/hr=1 -POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=0).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=0).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=1).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=1).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='0') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@test_table_n0 -POSTHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='0') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@test_table_n0 -# col_name data_type comment -key int -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 0] -Database: default -Table: test_table_n0 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 - numRows 670 - rawDataSize 5888 - totalSize 6558 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 -PREHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='1') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@test_table_n0 -POSTHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='1') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@test_table_n0 -# col_name data_type comment -key int -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 1] -Database: default -Table: test_table_n0 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 - numRows 330 - rawDataSize 2924 - totalSize 3254 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Storage Desc Params: - serialization.format 1 diff --git ql/src/test/results/clientpositive/infer_join_preds.q.out ql/src/test/results/clientpositive/infer_join_preds.q.out deleted file mode 100644 index 7aa1022e4d..0000000000 --- ql/src/test/results/clientpositive/infer_join_preds.q.out +++ /dev/null @@ -1,1328 +0,0 @@ -PREHOOK: query: explain -select * from src a join src1 b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from src a join src1 b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from src a join src1 b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from src a join src1 b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -128 val_128 128 -128 val_128 128 -128 val_128 128 -146 val_146 146 val_146 -146 val_146 146 val_146 -150 val_150 150 val_150 -213 val_213 213 val_213 -213 val_213 213 val_213 -224 val_224 224 -224 val_224 224 -238 val_238 238 val_238 -238 val_238 238 val_238 -255 val_255 255 val_255 -255 val_255 255 val_255 -273 val_273 273 val_273 -273 val_273 273 val_273 -273 val_273 273 val_273 -278 val_278 278 val_278 -278 val_278 278 val_278 -311 val_311 311 val_311 -311 val_311 311 val_311 -311 val_311 311 val_311 -369 val_369 369 -369 val_369 369 -369 val_369 369 -401 val_401 401 val_401 -401 val_401 401 val_401 -401 val_401 401 val_401 -401 val_401 401 val_401 -401 val_401 401 val_401 -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -66 val_66 66 val_66 -98 val_98 98 val_98 -98 val_98 98 val_98 -PREHOOK: query: explain -select * from -(select * from src where 1 = 0)a -join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -(select * from src where 1 = 0)a -join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 353 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 353 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from -(select * from src where 1 = 0)a -join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from -(select * from src where 1 = 0)a -join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -PREHOOK: query: explain -select * from -(select * from src where 1 = 0)a -left outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -(select * from src where 1 = 0)a -left outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from -(select * from src where 1 = 0)a -left outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from -(select * from src where 1 = 0)a -left outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -PREHOOK: query: explain -select * from -(select * from src where 1 = 0)a -right outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -(select * from src where 1 = 0)a -right outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 26 Data size: 4906 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 26 Data size: 4906 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from -(select * from src where 1 = 0)a -right outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from -(select * from src where 1 = 0)a -right outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL -NULL NULL -NULL NULL -NULL NULL -NULL NULL val_165 -NULL NULL val_193 -NULL NULL val_265 -NULL NULL val_27 -NULL NULL val_409 -NULL NULL val_484 -NULL NULL 128 -NULL NULL 146 val_146 -NULL NULL 150 val_150 -NULL NULL 213 val_213 -NULL NULL 224 -NULL NULL 238 val_238 -NULL NULL 255 val_255 -NULL NULL 273 val_273 -NULL NULL 278 val_278 -NULL NULL 311 val_311 -NULL NULL 369 -NULL NULL 401 val_401 -NULL NULL 406 val_406 -NULL NULL 66 val_66 -NULL NULL 98 val_98 -PREHOOK: query: explain -select * from -(select * from src where 1 = 0)a -full outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -(select * from src where 1 = 0)a -full outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from -(select * from src where 1 = 0)a -full outer join -(select * from src1)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from -(select * from src where 1 = 0)a -full outer join -(select * from src1)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL -NULL NULL -NULL NULL -NULL NULL -NULL NULL val_165 -NULL NULL val_193 -NULL NULL val_265 -NULL NULL val_27 -NULL NULL val_409 -NULL NULL val_484 -NULL NULL 128 -NULL NULL 146 val_146 -NULL NULL 150 val_150 -NULL NULL 213 val_213 -NULL NULL 224 -NULL NULL 238 val_238 -NULL NULL 255 val_255 -NULL NULL 273 val_273 -NULL NULL 278 val_278 -NULL NULL 311 val_311 -NULL NULL 369 -NULL NULL 401 val_401 -NULL NULL 406 val_406 -NULL NULL 66 val_66 -NULL NULL 98 val_98 -PREHOOK: query: explain -select * from -(select * from src)a -right outer join -(select * from src1 where 1 = 0)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -(select * from src)a -right outer join -(select * from src1 where 1 = 0)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from -(select * from src)a -right outer join -(select * from src1 where 1 = 0)b on a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: select * from -(select * from src)a -right outer join -(select * from src1 where 1 = 0)b on a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -PREHOOK: query: explain select * from src join src1 on src.key = src1.key and src.value = src1.value - where 4 between src.key and src.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from src join src1 on src.key = src1.key and src.value = src1.value - where 4 between src.key and src.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: 4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: 4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src1 - filterExpr: (4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) and key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) and key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col1 (type: string) - 1 _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: CREATE TABLE `table1_n8`( - `idp_warehouse_id` bigint, - `idp_audit_id` bigint, - `idp_effective_date` date, - `idp_end_date` date, - `idp_delete_date` date, - `pruid` varchar(32), - `prid` bigint, - `prtimesheetid` bigint, - `prassignmentid` bigint, - `prchargecodeid` bigint, - `prtypecodeid` bigint, - `prsequence` bigint, - `prmodby` varchar(96), - `prmodtime` timestamp, - `prrmexported` bigint, - `prrmckdel` bigint, - `slice_status` int, - `role_id` bigint, - `user_lov1` varchar(30), - `user_lov2` varchar(30), - `incident_id` bigint, - `incident_investment_id` bigint, - `odf_ss_actuals` bigint, - `practsum` decimal(38,20)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@table1_n8 -POSTHOOK: query: CREATE TABLE `table1_n8`( - `idp_warehouse_id` bigint, - `idp_audit_id` bigint, - `idp_effective_date` date, - `idp_end_date` date, - `idp_delete_date` date, - `pruid` varchar(32), - `prid` bigint, - `prtimesheetid` bigint, - `prassignmentid` bigint, - `prchargecodeid` bigint, - `prtypecodeid` bigint, - `prsequence` bigint, - `prmodby` varchar(96), - `prmodtime` timestamp, - `prrmexported` bigint, - `prrmckdel` bigint, - `slice_status` int, - `role_id` bigint, - `user_lov1` varchar(30), - `user_lov2` varchar(30), - `incident_id` bigint, - `incident_investment_id` bigint, - `odf_ss_actuals` bigint, - `practsum` decimal(38,20)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@table1_n8 -PREHOOK: query: CREATE TABLE `table2_n4`( - `idp_warehouse_id` bigint, - `idp_audit_id` bigint, - `idp_effective_date` date, - `idp_end_date` date, - `idp_delete_date` date, - `pruid` varchar(32), - `prid` bigint, - `prtimesheetid` bigint, - `prassignmentid` bigint, - `prchargecodeid` bigint, - `prtypecodeid` bigint, - `prsequence` bigint, - `prmodby` varchar(96), - `prmodtime` timestamp, - `prrmexported` bigint, - `prrmckdel` bigint, - `slice_status` int, - `role_id` bigint, - `user_lov1` varchar(30), - `user_lov2` varchar(30), - `incident_id` bigint, - `incident_investment_id` bigint, - `odf_ss_actuals` bigint, - `practsum` decimal(38,20)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@table2_n4 -POSTHOOK: query: CREATE TABLE `table2_n4`( - `idp_warehouse_id` bigint, - `idp_audit_id` bigint, - `idp_effective_date` date, - `idp_end_date` date, - `idp_delete_date` date, - `pruid` varchar(32), - `prid` bigint, - `prtimesheetid` bigint, - `prassignmentid` bigint, - `prchargecodeid` bigint, - `prtypecodeid` bigint, - `prsequence` bigint, - `prmodby` varchar(96), - `prmodtime` timestamp, - `prrmexported` bigint, - `prrmckdel` bigint, - `slice_status` int, - `role_id` bigint, - `user_lov1` varchar(30), - `user_lov2` varchar(30), - `incident_id` bigint, - `incident_investment_id` bigint, - `odf_ss_actuals` bigint, - `practsum` decimal(38,20)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@table2_n4 -PREHOOK: query: explain SELECT s.idp_warehouse_id AS source_warehouse_id - FROM table1_n8 s - JOIN - - table2_n4 d - ON ( - s.prid = d.prid ) - JOIN - table2_n4 e - ON - s.prid = e.prid - WHERE - concat( - CASE - WHEN s.prid IS NULL THEN 1 - ELSE s.prid - END,',', - CASE - WHEN s.prtimesheetid IS NULL THEN 1 - ELSE s.prtimesheetid - END,',', - CASE - WHEN s.prassignmentid IS NULL THEN 1 - ELSE s.prassignmentid - END,',', - CASE - WHEN s.prchargecodeid IS NULL THEN 1 - ELSE s.prchargecodeid - END,',', - CASE - WHEN (s.prtypecodeid) IS NULL THEN '' - ELSE s.prtypecodeid - END,',', - CASE - WHEN s.practsum IS NULL THEN 1 - ELSE s.practsum - END,',', - CASE - WHEN s.prsequence IS NULL THEN 1 - ELSE s.prsequence - END,',', - CASE - WHEN length(s.prmodby) IS NULL THEN '' - ELSE s.prmodby - END,',', - CASE - WHEN s.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) - ELSE s.prmodtime - END,',', - CASE - WHEN s.prrmexported IS NULL THEN 1 - ELSE s.prrmexported - END,',', - CASE - WHEN s.prrmckdel IS NULL THEN 1 - ELSE s.prrmckdel - END,',', - CASE - WHEN s.slice_status IS NULL THEN 1 - ELSE s.slice_status - END,',', - CASE - WHEN s.role_id IS NULL THEN 1 - ELSE s.role_id - END,',', - CASE - WHEN length(s.user_lov1) IS NULL THEN '' - ELSE s.user_lov1 - END,',', - CASE - WHEN length(s.user_lov2) IS NULL THEN '' - ELSE s.user_lov2 - END,',', - CASE - WHEN s.incident_id IS NULL THEN 1 - ELSE s.incident_id - END,',', - CASE - WHEN s.incident_investment_id IS NULL THEN 1 - ELSE s.incident_investment_id - END,',', - CASE - WHEN s.odf_ss_actuals IS NULL THEN 1 - ELSE s.odf_ss_actuals - END ) != concat( - CASE - WHEN length(d.pruid) IS NULL THEN '' - ELSE d.pruid - END,',', - CASE - WHEN d.prid IS NULL THEN 1 - ELSE d.prid - END,',', - CASE - WHEN d.prtimesheetid IS NULL THEN 1 - ELSE d.prtimesheetid - END,',', - CASE - WHEN d.prassignmentid IS NULL THEN 1 - ELSE d.prassignmentid - END,',', - CASE - WHEN d.prchargecodeid IS NULL THEN 1 - ELSE d.prchargecodeid - END,',', - CASE - WHEN (d.prtypecodeid) IS NULL THEN '' - ELSE d.prtypecodeid - END,',', - CASE - WHEN d.practsum IS NULL THEN 1 - ELSE d.practsum - END,',', - CASE - WHEN d.prsequence IS NULL THEN 1 - ELSE d.prsequence - END,',', - CASE - WHEN length(d.prmodby) IS NULL THEN '' - ELSE d.prmodby - END,',', - CASE - WHEN d.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) - ELSE d.prmodtime - END,',', - CASE - WHEN d.prrmexported IS NULL THEN 1 - ELSE d.prrmexported - END,',', - CASE - WHEN d.prrmckdel IS NULL THEN 1 - ELSE d.prrmckdel - END,',', - CASE - WHEN d.slice_status IS NULL THEN 1 - ELSE d.slice_status - END,',', - CASE - WHEN d.role_id IS NULL THEN 1 - ELSE d.role_id - END,',', - CASE - WHEN length(d.user_lov1) IS NULL THEN '' - ELSE d.user_lov1 - END,',', - CASE - WHEN length(d.user_lov2) IS NULL THEN '' - ELSE d.user_lov2 - END,',', - CASE - WHEN d.incident_id IS NULL THEN 1 - ELSE d.incident_id - END,',', - CASE - WHEN d.incident_investment_id IS NULL THEN 1 - ELSE d.incident_investment_id - END,',', - CASE - WHEN d.odf_ss_actuals IS NULL THEN 1 - ELSE d.odf_ss_actuals - END ) -PREHOOK: type: QUERY -PREHOOK: Input: default@table1_n8 -PREHOOK: Input: default@table2_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain SELECT s.idp_warehouse_id AS source_warehouse_id - FROM table1_n8 s - JOIN - - table2_n4 d - ON ( - s.prid = d.prid ) - JOIN - table2_n4 e - ON - s.prid = e.prid - WHERE - concat( - CASE - WHEN s.prid IS NULL THEN 1 - ELSE s.prid - END,',', - CASE - WHEN s.prtimesheetid IS NULL THEN 1 - ELSE s.prtimesheetid - END,',', - CASE - WHEN s.prassignmentid IS NULL THEN 1 - ELSE s.prassignmentid - END,',', - CASE - WHEN s.prchargecodeid IS NULL THEN 1 - ELSE s.prchargecodeid - END,',', - CASE - WHEN (s.prtypecodeid) IS NULL THEN '' - ELSE s.prtypecodeid - END,',', - CASE - WHEN s.practsum IS NULL THEN 1 - ELSE s.practsum - END,',', - CASE - WHEN s.prsequence IS NULL THEN 1 - ELSE s.prsequence - END,',', - CASE - WHEN length(s.prmodby) IS NULL THEN '' - ELSE s.prmodby - END,',', - CASE - WHEN s.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) - ELSE s.prmodtime - END,',', - CASE - WHEN s.prrmexported IS NULL THEN 1 - ELSE s.prrmexported - END,',', - CASE - WHEN s.prrmckdel IS NULL THEN 1 - ELSE s.prrmckdel - END,',', - CASE - WHEN s.slice_status IS NULL THEN 1 - ELSE s.slice_status - END,',', - CASE - WHEN s.role_id IS NULL THEN 1 - ELSE s.role_id - END,',', - CASE - WHEN length(s.user_lov1) IS NULL THEN '' - ELSE s.user_lov1 - END,',', - CASE - WHEN length(s.user_lov2) IS NULL THEN '' - ELSE s.user_lov2 - END,',', - CASE - WHEN s.incident_id IS NULL THEN 1 - ELSE s.incident_id - END,',', - CASE - WHEN s.incident_investment_id IS NULL THEN 1 - ELSE s.incident_investment_id - END,',', - CASE - WHEN s.odf_ss_actuals IS NULL THEN 1 - ELSE s.odf_ss_actuals - END ) != concat( - CASE - WHEN length(d.pruid) IS NULL THEN '' - ELSE d.pruid - END,',', - CASE - WHEN d.prid IS NULL THEN 1 - ELSE d.prid - END,',', - CASE - WHEN d.prtimesheetid IS NULL THEN 1 - ELSE d.prtimesheetid - END,',', - CASE - WHEN d.prassignmentid IS NULL THEN 1 - ELSE d.prassignmentid - END,',', - CASE - WHEN d.prchargecodeid IS NULL THEN 1 - ELSE d.prchargecodeid - END,',', - CASE - WHEN (d.prtypecodeid) IS NULL THEN '' - ELSE d.prtypecodeid - END,',', - CASE - WHEN d.practsum IS NULL THEN 1 - ELSE d.practsum - END,',', - CASE - WHEN d.prsequence IS NULL THEN 1 - ELSE d.prsequence - END,',', - CASE - WHEN length(d.prmodby) IS NULL THEN '' - ELSE d.prmodby - END,',', - CASE - WHEN d.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) - ELSE d.prmodtime - END,',', - CASE - WHEN d.prrmexported IS NULL THEN 1 - ELSE d.prrmexported - END,',', - CASE - WHEN d.prrmckdel IS NULL THEN 1 - ELSE d.prrmckdel - END,',', - CASE - WHEN d.slice_status IS NULL THEN 1 - ELSE d.slice_status - END,',', - CASE - WHEN d.role_id IS NULL THEN 1 - ELSE d.role_id - END,',', - CASE - WHEN length(d.user_lov1) IS NULL THEN '' - ELSE d.user_lov1 - END,',', - CASE - WHEN length(d.user_lov2) IS NULL THEN '' - ELSE d.user_lov2 - END,',', - CASE - WHEN d.incident_id IS NULL THEN 1 - ELSE d.incident_id - END,',', - CASE - WHEN d.incident_investment_id IS NULL THEN 1 - ELSE d.incident_investment_id - END,',', - CASE - WHEN d.odf_ss_actuals IS NULL THEN 1 - ELSE d.odf_ss_actuals - END ) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@table1_n8 -POSTHOOK: Input: default@table2_n4 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s - filterExpr: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: idp_warehouse_id (type: bigint), prid (type: bigint), concat(CAST( CASE WHEN (prid is null) THEN (1L) ELSE (prid) END AS STRING), ',', CASE WHEN (prtimesheetid is null) THEN (1L) ELSE (prtimesheetid) END, ',', CASE WHEN (prassignmentid is null) THEN (1L) ELSE (prassignmentid) END, ',', CASE WHEN (prchargecodeid is null) THEN (1L) ELSE (prchargecodeid) END, ',', CASE WHEN (prtypecodeid is null) THEN ('') ELSE (CAST( prtypecodeid AS STRING)) END, ',', CASE WHEN (practsum is null) THEN (1) ELSE (practsum) END, ',', CASE WHEN (prsequence is null) THEN (1L) ELSE (prsequence) END, ',', CASE WHEN (length(prmodby) is null) THEN ('') ELSE (prmodby) END, ',', CASE WHEN (prmodtime is null) THEN (TIMESTAMP'2017-12-08 00:00:00') ELSE (prmodtime) END, ',', CASE WHEN (prrmexported is null) THEN (1L) ELSE (prrmexported) END, ',', CASE WHEN (prrmckdel is null) THEN (1L) ELSE (prrmckdel) END, ',', CASE WHEN (slice_status is null) THEN (1) ELSE (slice_status) END, ',', CASE WHEN (role_id is null) THEN (1L) ELSE (role_id) END, ',', CASE WHEN (length(user_lov1) is null) THEN ('') ELSE (user_lov1) END, ',', CASE WHEN (length(user_lov2) is null) THEN ('') ELSE (user_lov2) END, ',', CASE WHEN (incident_id is null) THEN (1L) ELSE (incident_id) END, ',', CASE WHEN (incident_investment_id is null) THEN (1L) ELSE (incident_investment_id) END, ',', CASE WHEN (odf_ss_actuals is null) THEN (1L) ELSE (odf_ss_actuals) END) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: bigint) - Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: bigint), _col2 (type: string) - TableScan - alias: e - filterExpr: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: prid (type: bigint) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: bigint) - 1 _col0 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 734 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: bigint) - Statistics: Num rows: 1 Data size: 734 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: bigint), _col2 (type: string) - TableScan - alias: d - filterExpr: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: prid is not null (type: boolean) - Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: prid (type: bigint), concat(CASE WHEN (length(pruid) is null) THEN ('') ELSE (pruid) END, ',', CASE WHEN (prid is null) THEN (1L) ELSE (prid) END, ',', CASE WHEN (prtimesheetid is null) THEN (1L) ELSE (prtimesheetid) END, ',', CASE WHEN (prassignmentid is null) THEN (1L) ELSE (prassignmentid) END, ',', CASE WHEN (prchargecodeid is null) THEN (1L) ELSE (prchargecodeid) END, ',', CASE WHEN (prtypecodeid is null) THEN ('') ELSE (CAST( prtypecodeid AS STRING)) END, ',', CASE WHEN (practsum is null) THEN (1) ELSE (practsum) END, ',', CASE WHEN (prsequence is null) THEN (1L) ELSE (prsequence) END, ',', CASE WHEN (length(prmodby) is null) THEN ('') ELSE (prmodby) END, ',', CASE WHEN (prmodtime is null) THEN (TIMESTAMP'2017-12-08 00:00:00') ELSE (prmodtime) END, ',', CASE WHEN (prrmexported is null) THEN (1L) ELSE (prrmexported) END, ',', CASE WHEN (prrmckdel is null) THEN (1L) ELSE (prrmckdel) END, ',', CASE WHEN (slice_status is null) THEN (1) ELSE (slice_status) END, ',', CASE WHEN (role_id is null) THEN (1L) ELSE (role_id) END, ',', CASE WHEN (length(user_lov1) is null) THEN ('') ELSE (user_lov1) END, ',', CASE WHEN (length(user_lov2) is null) THEN ('') ELSE (user_lov2) END, ',', CASE WHEN (incident_id is null) THEN (1L) ELSE (incident_id) END, ',', CASE WHEN (incident_investment_id is null) THEN (1L) ELSE (incident_investment_id) END, ',', CASE WHEN (odf_ss_actuals is null) THEN (1L) ELSE (odf_ss_actuals) END) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: bigint) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: bigint) - Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: bigint) - 1 _col0 (type: bigint) - outputColumnNames: _col0, _col2, _col5 - Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (_col2 <> _col5) (type: boolean) - Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: bigint) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: drop table table2_n4 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@table2_n4 -PREHOOK: Output: default@table2_n4 -POSTHOOK: query: drop table table2_n4 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@table2_n4 -POSTHOOK: Output: default@table2_n4 -PREHOOK: query: drop table table1_n8 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@table1_n8 -PREHOOK: Output: default@table1_n8 -POSTHOOK: query: drop table table1_n8 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@table1_n8 -POSTHOOK: Output: default@table1_n8 diff --git ql/src/test/results/clientpositive/innerjoin1.q.out ql/src/test/results/clientpositive/innerjoin1.q.out deleted file mode 100644 index 8617015955..0000000000 --- ql/src/test/results/clientpositive/innerjoin1.q.out +++ /dev/null @@ -1,250 +0,0 @@ -PREHOOK: query: DROP TABLE IF EXISTS innerjoinsrc1 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE IF EXISTS innerjoinsrc1 -POSTHOOK: type: DROPTABLE -PREHOOK: query: DROP TABLE IF EXISTS innerjoinsrc2 -PREHOOK: type: DROPTABLE -POSTHOOK: query: DROP TABLE IF EXISTS innerjoinsrc2 -POSTHOOK: type: DROPTABLE -PREHOOK: query: CREATE TABLE innerjoinsrc1 (a1 STRING, a2 STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@innerjoinsrc1 -POSTHOOK: query: CREATE TABLE innerjoinsrc1 (a1 STRING, a2 STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@innerjoinsrc1 -PREHOOK: query: CREATE TABLE innerjoinsrc2 (b1 STRING, b2 STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@innerjoinsrc2 -POSTHOOK: query: CREATE TABLE innerjoinsrc2 (b1 STRING, b2 STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@innerjoinsrc2 -PREHOOK: query: INSERT INTO TABLE innerjoinsrc1 (a1) VALUES ('1') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@innerjoinsrc1 -POSTHOOK: query: INSERT INTO TABLE innerjoinsrc1 (a1) VALUES ('1') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@innerjoinsrc1 -POSTHOOK: Lineage: innerjoinsrc1.a1 SCRIPT [] -POSTHOOK: Lineage: innerjoinsrc1.a2 SIMPLE [] -PREHOOK: query: INSERT INTO TABLE innerjoinsrc1 VALUES ('2', '2') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@innerjoinsrc1 -POSTHOOK: query: INSERT INTO TABLE innerjoinsrc1 VALUES ('2', '2') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@innerjoinsrc1 -POSTHOOK: Lineage: innerjoinsrc1.a1 SCRIPT [] -POSTHOOK: Lineage: innerjoinsrc1.a2 SCRIPT [] -PREHOOK: query: INSERT INTO TABLE innerjoinsrc2 (b1) VALUES ('1') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@innerjoinsrc2 -POSTHOOK: query: INSERT INTO TABLE innerjoinsrc2 (b1) VALUES ('1') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@innerjoinsrc2 -POSTHOOK: Lineage: innerjoinsrc2.b1 SCRIPT [] -POSTHOOK: Lineage: innerjoinsrc2.b2 SIMPLE [] -PREHOOK: query: INSERT INTO TABLE innerjoinsrc2 VALUES ('2', '2') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@innerjoinsrc2 -POSTHOOK: query: INSERT INTO TABLE innerjoinsrc2 VALUES ('2', '2') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@innerjoinsrc2 -POSTHOOK: Lineage: innerjoinsrc2.b1 SCRIPT [] -POSTHOOK: Lineage: innerjoinsrc2.b2 SCRIPT [] -PREHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) -PREHOOK: type: QUERY -PREHOOK: Input: default@innerjoinsrc1 -PREHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@innerjoinsrc1 -POSTHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: c1 - filterExpr: (a1 is not null or a2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (a1 is not null or a2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: a1 (type: string), a2 (type: string), CASE WHEN (a1 is not null) THEN (a1) ELSE (a2) END (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: c2 - filterExpr: (b1 is not null or b2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (b1 is not null or b2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: b1 (type: string), b2 (type: string), CASE WHEN (b1 is not null) THEN (b1) ELSE (b2) END (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: string) - 1 _col2 (type: string) - outputColumnNames: _col0, _col1, _col3, _col4 - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) -PREHOOK: type: QUERY -PREHOOK: Input: default@innerjoinsrc1 -PREHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@innerjoinsrc1 -POSTHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -1 NULL 1 NULL -2 2 2 2 -PREHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) -PREHOOK: type: QUERY -PREHOOK: Input: default@innerjoinsrc1 -PREHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@innerjoinsrc1 -POSTHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: c1 - filterExpr: (a1 is not null or a2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (a1 is not null or a2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: a1 (type: string), a2 (type: string), CASE WHEN (a1 is not null) THEN (a1) ELSE (a2) END (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: c2 - filterExpr: (b1 is not null or b2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (b1 is not null or b2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: b1 (type: string), b2 (type: string), CASE WHEN (b1 is not null) THEN (b1) ELSE (b2) END (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: string) - 1 _col2 (type: string) - outputColumnNames: _col0, _col1, _col3, _col4 - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) -PREHOOK: type: QUERY -PREHOOK: Input: default@innerjoinsrc1 -PREHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@innerjoinsrc1 -POSTHOOK: Input: default@innerjoinsrc2 -#### A masked pattern was here #### -1 NULL 1 NULL -2 2 2 2 diff --git ql/src/test/results/clientpositive/input11.q.out ql/src/test/results/clientpositive/input11.q.out deleted file mode 100644 index 5d2a1e19bc..0000000000 --- ql/src/test/results/clientpositive/input11.q.out +++ /dev/null @@ -1,243 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n152(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n152 -POSTHOOK: query: CREATE TABLE dest1_n152(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n152 -PREHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n152 -POSTHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n152 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n152 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n152 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.dest1_n152 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n152 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n152 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n152 -POSTHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n152 -POSTHOOK: Lineage: dest1_n152.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n152.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n152.* FROM dest1_n152 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n152 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n152.* FROM dest1_n152 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n152 -#### A masked pattern was here #### -86 val_86 -27 val_27 -98 val_98 -66 val_66 -37 val_37 -15 val_15 -82 val_82 -17 val_17 -0 val_0 -57 val_57 -20 val_20 -92 val_92 -47 val_47 -72 val_72 -4 val_4 -35 val_35 -54 val_54 -51 val_51 -65 val_65 -83 val_83 -12 val_12 -67 val_67 -84 val_84 -58 val_58 -8 val_8 -24 val_24 -42 val_42 -0 val_0 -96 val_96 -26 val_26 -51 val_51 -43 val_43 -95 val_95 -98 val_98 -85 val_85 -77 val_77 -0 val_0 -87 val_87 -15 val_15 -72 val_72 -90 val_90 -19 val_19 -10 val_10 -5 val_5 -58 val_58 -35 val_35 -95 val_95 -11 val_11 -34 val_34 -42 val_42 -78 val_78 -76 val_76 -41 val_41 -30 val_30 -64 val_64 -76 val_76 -74 val_74 -69 val_69 -33 val_33 -70 val_70 -5 val_5 -2 val_2 -35 val_35 -80 val_80 -44 val_44 -53 val_53 -90 val_90 -12 val_12 -5 val_5 -70 val_70 -24 val_24 -70 val_70 -83 val_83 -26 val_26 -67 val_67 -18 val_18 -9 val_9 -18 val_18 -97 val_97 -84 val_84 -28 val_28 -37 val_37 -90 val_90 -97 val_97 diff --git ql/src/test/results/clientpositive/input11_limit.q.out ql/src/test/results/clientpositive/input11_limit.q.out deleted file mode 100644 index 60f5709091..0000000000 --- ql/src/test/results/clientpositive/input11_limit.q.out +++ /dev/null @@ -1,115 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n153(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n153 -POSTHOOK: query: CREATE TABLE dest1_n153(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n153 -PREHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n153 -POSTHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n153 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n153 - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n153 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - -PREHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n153 -POSTHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n153 -POSTHOOK: Lineage: dest1_n153.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n153.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n153.* FROM dest1_n153 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n153 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n153.* FROM dest1_n153 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n153 -#### A masked pattern was here #### -0 val_0 -15 val_15 -17 val_17 -27 val_27 -37 val_37 -57 val_57 -66 val_66 -82 val_82 -86 val_86 -98 val_98 diff --git ql/src/test/results/clientpositive/input1_limit.q.out ql/src/test/results/clientpositive/input1_limit.q.out deleted file mode 100644 index b2e87859b7..0000000000 --- ql/src/test/results/clientpositive/input1_limit.q.out +++ /dev/null @@ -1,300 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n12(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n12 -POSTHOOK: query: CREATE TABLE dest1_n12(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n12 -PREHOOK: query: CREATE TABLE dest2_n2(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest2_n2 -POSTHOOK: query: CREATE TABLE dest2_n2(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest2_n2 -PREHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n12 -PREHOOK: Output: default@dest2_n2 -POSTHOOK: query: EXPLAIN -FROM src -INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n12 -POSTHOOK: Output: default@dest2_n2 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 100) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string) - Filter Operator - predicate: (key < 100) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n12 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n12 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.dest1_n12 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.dest2_n2 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n2 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n2 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n12 -PREHOOK: Output: default@dest2_n2 -POSTHOOK: query: FROM src -INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 -INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n12 -POSTHOOK: Output: default@dest2_n2 -POSTHOOK: Lineage: dest1_n12.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n12.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n12.* FROM dest1_n12 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n12 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n12.* FROM dest1_n12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n12 -#### A masked pattern was here #### -0 val_0 -15 val_15 -17 val_17 -27 val_27 -37 val_37 -57 val_57 -66 val_66 -82 val_82 -86 val_86 -98 val_98 -PREHOOK: query: SELECT dest2_n2.* FROM dest2_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n2 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest2_n2.* FROM dest2_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n2 -#### A masked pattern was here #### -27 val_27 -37 val_37 -66 val_66 -86 val_86 -98 val_98 diff --git ql/src/test/results/clientpositive/input21.q.out ql/src/test/results/clientpositive/input21.q.out deleted file mode 100644 index c5481c7797..0000000000 --- ql/src/test/results/clientpositive/input21.q.out +++ /dev/null @@ -1,84 +0,0 @@ -PREHOOK: query: CREATE TABLE src_null(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@src_null -POSTHOOK: query: CREATE TABLE src_null(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@src_null -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE src_null -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@src_null -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE src_null -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@src_null -PREHOOK: query: EXPLAIN SELECT * FROM src_null DISTRIBUTE BY c SORT BY d -PREHOOK: type: QUERY -PREHOOK: Input: default@src_null -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src_null DISTRIBUTE BY c SORT BY d -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_null -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_null - Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string), c (type: string), d (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col3 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string), KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM src_null DISTRIBUTE BY c SORT BY d -PREHOOK: type: QUERY -PREHOOK: Input: default@src_null -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM src_null DISTRIBUTE BY c SORT BY d -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_null -#### A masked pattern was here #### -1.0 1 same 0 -1.0 1 same 1 -1.0 1 same 2 -1.0 1 same 3 -1.0 1 same 4 -NULL 1 same 5 -NULL NULL same 6 -1.0 NULL same 7 -1.0 1 same 8 -1.0 1 same 9 diff --git ql/src/test/results/clientpositive/input22.q.out ql/src/test/results/clientpositive/input22.q.out deleted file mode 100644 index a6dbe370e6..0000000000 --- ql/src/test/results/clientpositive/input22.q.out +++ /dev/null @@ -1,100 +0,0 @@ -PREHOOK: query: CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@INPUT4 -POSTHOOK: query: CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@INPUT4 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE INPUT4 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@input4 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE INPUT4 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@input4 -PREHOOK: query: EXPLAIN -SELECT a.KEY2 -FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 - FROM INPUT4) a -ORDER BY KEY2 LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@input4 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT a.KEY2 -FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 - FROM INPUT4) a -ORDER BY KEY2 LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@input4 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: input4 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT a.KEY2 -FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 - FROM INPUT4) a -ORDER BY KEY2 LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@input4 -#### A masked pattern was here #### -POSTHOOK: query: SELECT a.KEY2 -FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 - FROM INPUT4) a -ORDER BY KEY2 LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@input4 -#### A masked pattern was here #### -0 -0 -0 -10 -100 -100 -103 -103 -104 -104 diff --git ql/src/test/results/clientpositive/input23.q.out ql/src/test/results/clientpositive/input23.q.out deleted file mode 100644 index cd79dd38c1..0000000000 --- ql/src/test/results/clientpositive/input23.q.out +++ /dev/null @@ -1,182 +0,0 @@ -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain extended - select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: explain extended - select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, CAST('2008-04-08' AS STRING) AS `ds`, CAST('11' AS STRING) AS `hr`, `t2`.`key` AS `key1`, `t2`.`value` AS `value1`, CAST('2008-04-08' AS STRING) AS `ds1`, CAST('14' AS STRING) AS `hr1` -FROM (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' AND `hr` = '11') AS `t0`, -(SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' AND `hr` = '14') AS `t2` -LIMIT 5 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((ds = '2008-04-08') and (hr = '11')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col0 (type: string), _col1 (type: string) - auto parallelism: false - TableScan - alias: b - filterExpr: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col0 (type: string), _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 173000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string), _col2 (type: string), _col3 (type: string), '2008-04-08' (type: string), '14' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 500 Data size: 353000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 5 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 - columns.types string:string:string:string:string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: 5 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input25.q.out ql/src/test/results/clientpositive/input25.q.out deleted file mode 100644 index efc337c9b3..0000000000 --- ql/src/test/results/clientpositive/input25.q.out +++ /dev/null @@ -1,181 +0,0 @@ -PREHOOK: query: create table tst(a int, b int) partitioned by (d string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tst -POSTHOOK: query: create table tst(a int, b int) partitioned by (d string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tst -PREHOOK: query: alter table tst add partition (d='2009-01-01') -PREHOOK: type: ALTERTABLE_ADDPARTS -PREHOOK: Output: default@tst -POSTHOOK: query: alter table tst add partition (d='2009-01-01') -POSTHOOK: type: ALTERTABLE_ADDPARTS -POSTHOOK: Output: default@tst -POSTHOOK: Output: default@tst@d=2009-01-01 -PREHOOK: query: alter table tst add partition (d='2009-02-02') -PREHOOK: type: ALTERTABLE_ADDPARTS -PREHOOK: Output: default@tst -POSTHOOK: query: alter table tst add partition (d='2009-02-02') -POSTHOOK: type: ALTERTABLE_ADDPARTS -POSTHOOK: Output: default@tst -POSTHOOK: Output: default@tst@d=2009-02-02 -PREHOOK: query: explain -select * from ( - select * from (select * from tst x where x.d='2009-01-01' limit 10)a - union all - select * from (select * from tst x where x.d='2009-02-02' limit 10)b -) subq -PREHOOK: type: QUERY -PREHOOK: Input: default@tst -PREHOOK: Input: default@tst@d=2009-01-01 -PREHOOK: Input: default@tst@d=2009-02-02 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from ( - select * from (select * from tst x where x.d='2009-01-01' limit 10)a - union all - select * from (select * from tst x where x.d='2009-02-02' limit 10)b -) subq -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tst -POSTHOOK: Input: default@tst@d=2009-01-01 -POSTHOOK: Input: default@tst@d=2009-02-02 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (d = '2009-01-01') (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: a (type: int), b (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: int), _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), '2009-01-01' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Union - Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TableScan - Union - Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (d = '2009-02-02') (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: a (type: int), b (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: int), _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), '2009-02-02' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from ( - select * from (select * from tst x where x.d='2009-01-01' limit 10)a - union all - select * from (select * from tst x where x.d='2009-02-02' limit 10)b -) subq -PREHOOK: type: QUERY -PREHOOK: Input: default@tst -PREHOOK: Input: default@tst@d=2009-01-01 -PREHOOK: Input: default@tst@d=2009-02-02 -#### A masked pattern was here #### -POSTHOOK: query: select * from ( - select * from (select * from tst x where x.d='2009-01-01' limit 10)a - union all - select * from (select * from tst x where x.d='2009-02-02' limit 10)b -) subq -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tst -POSTHOOK: Input: default@tst@d=2009-01-01 -POSTHOOK: Input: default@tst@d=2009-02-02 -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input26.q.out ql/src/test/results/clientpositive/input26.q.out deleted file mode 100644 index 82d6fb6ff9..0000000000 --- ql/src/test/results/clientpositive/input26.q.out +++ /dev/null @@ -1,169 +0,0 @@ -PREHOOK: query: explain -select * from ( - select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa - union all - select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb -)subq -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from ( - select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa - union all - select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb -)subq -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((ds = '2008-04-08') and (hr = '11')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Union - Statistics: Num rows: 6 Data size: 1574 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TableScan - Union - Statistics: Num rows: 6 Data size: 1574 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: b - filterExpr: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '14' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 254 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from ( - select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa - union all - select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb -)subq -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: select * from ( - select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa - union all - select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb -)subq -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -10 val_10 2008-04-08 11 -100 val_100 2008-04-08 11 diff --git ql/src/test/results/clientpositive/input2_limit.q.out ql/src/test/results/clientpositive/input2_limit.q.out deleted file mode 100644 index b099ffdf25..0000000000 --- ql/src/test/results/clientpositive/input2_limit.q.out +++ /dev/null @@ -1,60 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (UDFToDouble(key) < 300.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 300.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: 5 - Processor Tree: - ListSink - -PREHOOK: query: SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -238 val_238 -86 val_86 -27 val_27 -165 val_165 -255 val_255 diff --git ql/src/test/results/clientpositive/input30.q.out ql/src/test/results/clientpositive/input30.q.out deleted file mode 100644 index 84fbd17ef1..0000000000 --- ql/src/test/results/clientpositive/input30.q.out +++ /dev/null @@ -1,247 +0,0 @@ -PREHOOK: query: create table dest30(a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest30 -POSTHOOK: query: create table dest30(a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest30 -PREHOOK: query: create table tst_dest30(a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tst_dest30 -POSTHOOK: query: create table tst_dest30(a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tst_dest30 -PREHOOK: query: explain -insert overwrite table dest30 -select count(1) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tst_dest30 -POSTHOOK: query: explain -insert overwrite table dest30 -select count(1) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tst_dest30 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((hash(rand(460476415)) & 2147483647) % 32) = 0) (type: boolean) - Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tst_dest30 - Select Operator - expressions: _col0 (type: int) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(a, 'hll') - mode: complete - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: struct) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tst_dest30 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: a - Column Types: int - Table: default.tst_dest30 - -PREHOOK: query: insert overwrite table dest30 -select count(1) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tst_dest30 -POSTHOOK: query: insert overwrite table dest30 -select count(1) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tst_dest30 -POSTHOOK: Lineage: tst_dest30.a EXPRESSION [(src)src.null, ] -PREHOOK: query: select * from tst_dest30 -PREHOOK: type: QUERY -PREHOOK: Input: default@tst_dest30 -#### A masked pattern was here #### -POSTHOOK: query: select * from tst_dest30 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tst_dest30 -#### A masked pattern was here #### -15 -PREHOOK: query: explain -insert overwrite table dest30 -select count(1) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest30 -POSTHOOK: query: explain -insert overwrite table dest30 -select count(1) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest30 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: $f0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger($f0) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest30 - Select Operator - expressions: _col0 (type: int) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(a, 'hll') - mode: complete - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: struct) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest30 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: a - Column Types: int - Table: default.dest30 - -PREHOOK: query: insert overwrite table dest30 -select count(1) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest30 -POSTHOOK: query: insert overwrite table dest30 -select count(1) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest30 -POSTHOOK: Lineage: dest30.a EXPRESSION [] -PREHOOK: query: select * from tst_dest30 -PREHOOK: type: QUERY -PREHOOK: Input: default@tst_dest30 -#### A masked pattern was here #### -POSTHOOK: query: select * from tst_dest30 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tst_dest30 -#### A masked pattern was here #### -15 diff --git ql/src/test/results/clientpositive/input32.q.out ql/src/test/results/clientpositive/input32.q.out deleted file mode 100644 index 76c46e72e7..0000000000 --- ql/src/test/results/clientpositive/input32.q.out +++ /dev/null @@ -1,131 +0,0 @@ -PREHOOK: query: create table dest32(a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest32 -POSTHOOK: query: create table dest32(a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest32 -PREHOOK: query: create table tst_dest32(a int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tst_dest32 -POSTHOOK: query: create table tst_dest32(a int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tst_dest32 -PREHOOK: query: explain -insert overwrite table dest32 -select count(1) from srcbucket -PREHOOK: type: QUERY -PREHOOK: Input: default@srcbucket -PREHOOK: Output: default@tst_dest32 -POSTHOOK: query: explain -insert overwrite table dest32 -select count(1) from srcbucket -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: default@tst_dest32 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcbucket - Statistics: Num rows: 1000 Data size: 10603 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1000 Data size: 10603 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tst_dest32 - Select Operator - expressions: _col0 (type: int) - outputColumnNames: a - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(a, 'hll') - mode: complete - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: struct) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tst_dest32 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: a - Column Types: int - Table: default.tst_dest32 - -PREHOOK: query: insert overwrite table dest32 -select count(1) from srcbucket -PREHOOK: type: QUERY -PREHOOK: Input: default@srcbucket -PREHOOK: Output: default@tst_dest32 -POSTHOOK: query: insert overwrite table dest32 -select count(1) from srcbucket -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcbucket -POSTHOOK: Output: default@tst_dest32 -POSTHOOK: Lineage: tst_dest32.a EXPRESSION [(srcbucket)srcbucket.null, ] -PREHOOK: query: select * from tst_dest32 -PREHOOK: type: QUERY -PREHOOK: Input: default@tst_dest32 -#### A masked pattern was here #### -POSTHOOK: query: select * from tst_dest32 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tst_dest32 -#### A masked pattern was here #### -1000 diff --git ql/src/test/results/clientpositive/input3_limit.q.out ql/src/test/results/clientpositive/input3_limit.q.out deleted file mode 100644 index a39f8ccda5..0000000000 --- ql/src/test/results/clientpositive/input3_limit.q.out +++ /dev/null @@ -1,203 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n78(key STRING, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n78 -POSTHOOK: query: CREATE TABLE T1_n78(key STRING, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n78 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE T1_n78 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n78 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE T1_n78 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n78 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv2.txt' INTO TABLE T1_n78 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n78 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv2.txt' INTO TABLE T1_n78 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n78 -PREHOOK: query: CREATE TABLE T2_n47(key STRING, value STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n47 -POSTHOOK: query: CREATE TABLE T2_n47(key STRING, value STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n47 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n78 -PREHOOK: Output: default@t2_n47 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n78 -POSTHOOK: Output: default@t2_n47 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n78 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 20 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n47 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.t2_n47 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.t2_n47 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n78 -PREHOOK: Output: default@t2_n47 -POSTHOOK: query: INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n78 -POSTHOOK: Output: default@t2_n47 -POSTHOOK: Lineage: t2_n47.key SIMPLE [(t1_n78)t1_n78.FieldSchema(name:key, type:string, comment:null), ] -POSTHOOK: Lineage: t2_n47.value SIMPLE [(t1_n78)t1_n78.FieldSchema(name:value, type:string, comment:null), ] -PREHOOK: query: SELECT * FROM T2_n47 ORDER BY key, value -PREHOOK: type: QUERY -PREHOOK: Input: default@t2_n47 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM T2_n47 ORDER BY key, value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t2_n47 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -0 val_1 -0 val_1 -1 val_2 -10 val_10 -10 val_11 -100 val_100 -100 val_100 -100 val_101 -100 val_101 -101 val_102 -102 val_103 -103 val_103 -103 val_103 -104 val_104 -104 val_104 -104 val_105 -104 val_105 diff --git ql/src/test/results/clientpositive/input4_limit.q.out ql/src/test/results/clientpositive/input4_limit.q.out deleted file mode 100644 index 34f4821c7d..0000000000 --- ql/src/test/results/clientpositive/input4_limit.q.out +++ /dev/null @@ -1,101 +0,0 @@ -PREHOOK: query: explain -select * from src sort by key limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from src sort by key limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: select * from src sort by key limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from src sort by key limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -10 val_10 -100 val_100 -100 val_100 -103 val_103 -103 val_103 -104 val_104 -104 val_104 diff --git ql/src/test/results/clientpositive/input6.q.out ql/src/test/results/clientpositive/input6.q.out deleted file mode 100644 index 799634a460..0000000000 --- ql/src/test/results/clientpositive/input6.q.out +++ /dev/null @@ -1,159 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n35(key STRING, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n35 -POSTHOOK: query: CREATE TABLE dest1_n35(key STRING, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n35 -PREHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n35 -POSTHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n35 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is null (type: boolean) - Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: null (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n35 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n35 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.dest1_n35 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n35 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n35 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n35 -POSTHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n35 -POSTHOOK: Lineage: dest1_n35.key SIMPLE [] -POSTHOOK: Lineage: dest1_n35.value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n35.* FROM dest1_n35 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n35 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n35.* FROM dest1_n35 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n35 -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input7.q.out ql/src/test/results/clientpositive/input7.q.out deleted file mode 100644 index d6e2d14cba..0000000000 --- ql/src/test/results/clientpositive/input7.q.out +++ /dev/null @@ -1,180 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n167(c1 DOUBLE, c2 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n167 -POSTHOOK: query: CREATE TABLE dest1_n167(c1 DOUBLE, c2 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n167 -PREHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n167 -POSTHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n167 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: null (type: double), UDFToInteger(key) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n167 - Select Operator - expressions: _col0 (type: double), _col1 (type: int) - outputColumnNames: c1, c2 - Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n167 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2 - Column Types: double, int - Table: default.dest1_n167 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n167 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n167 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n167 -POSTHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n167 -POSTHOOK: Lineage: dest1_n167.c1 EXPRESSION [] -POSTHOOK: Lineage: dest1_n167.c2 EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n167.* FROM dest1_n167 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n167 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n167.* FROM dest1_n167 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n167 -#### A masked pattern was here #### -NULL 238 -NULL NULL -NULL 311 -NULL NULL -NULL NULL -NULL NULL -NULL 255 -NULL 278 -NULL 98 -NULL NULL -NULL NULL -NULL NULL -NULL 401 -NULL 150 -NULL 273 -NULL 224 -NULL 369 -NULL 66 -NULL 128 -NULL 213 -NULL 146 -NULL 406 -NULL NULL -NULL NULL -NULL NULL diff --git ql/src/test/results/clientpositive/input8.q.out ql/src/test/results/clientpositive/input8.q.out deleted file mode 100644 index d3034ca14e..0000000000 --- ql/src/test/results/clientpositive/input8.q.out +++ /dev/null @@ -1,181 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n28(c1 STRING, c2 INT, c3 DOUBLE) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n28 -POSTHOOK: query: CREATE TABLE dest1_n28(c1 STRING, c2 INT, c3 DOUBLE) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n28 -PREHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n28 -POSTHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n28 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: null (type: string), null (type: int), null (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n28 - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: double) - outputColumnNames: c1, c2, c3 - Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n28 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3 - Column Types: string, int, double - Table: default.dest1_n28 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n28 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n28 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n28 -POSTHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n28 -POSTHOOK: Lineage: dest1_n28.c1 EXPRESSION [] -POSTHOOK: Lineage: dest1_n28.c2 EXPRESSION [] -POSTHOOK: Lineage: dest1_n28.c3 EXPRESSION [] -PREHOOK: query: SELECT dest1_n28.* FROM dest1_n28 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n28 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n28.* FROM dest1_n28 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n28 -#### A masked pattern was here #### -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL -NULL NULL NULL diff --git ql/src/test/results/clientpositive/input9.q.out ql/src/test/results/clientpositive/input9.q.out deleted file mode 100644 index 1aafc79fd8..0000000000 --- ql/src/test/results/clientpositive/input9.q.out +++ /dev/null @@ -1,162 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n159(value STRING, key INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n159 -POSTHOOK: query: CREATE TABLE dest1_n159(value STRING, key INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n159 -PREHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n159 -POSTHOOK: query: EXPLAIN -FROM src1 -INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n159 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 86 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: null (type: string), UDFToInteger(_col0) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n159 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: value, key - Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(value, 'hll'), compute_stats(key, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n159 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: value, key - Column Types: string, int - Table: default.dest1_n159 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n159 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n159 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest1_n159 -POSTHOOK: query: FROM src1 -INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest1_n159 -POSTHOOK: Lineage: dest1_n159.key EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n159.value EXPRESSION [] -PREHOOK: query: SELECT dest1_n159.* FROM dest1_n159 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n159 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n159.* FROM dest1_n159 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n159 -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input_part1.q.out ql/src/test/results/clientpositive/input_part1.q.out deleted file mode 100644 index 3ac660202f..0000000000 --- ql/src/test/results/clientpositive/input_part1.q.out +++ /dev/null @@ -1,535 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n45(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n45 -POSTHOOK: query: CREATE TABLE dest1_n45(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n45 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart -INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n45 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart -INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n45 -OPTIMIZED SQL: SELECT `key`, `value`, CAST('12' AS STRING) AS `hr`, CAST('2008-04-08' AS STRING) AS `ds` -FROM `default`.`srcpart` -WHERE `key` < 100 AND `ds` = '2008-04-08' AND `hr` = '12' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08') and (hr = '12')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, hr, ds - Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, hr, ds - Column Types: int, string, string, string - Table: default.dest1_n45 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - name: default.dest1_n45 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n45 - name: default.dest1_n45 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM srcpart -INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@dest1_n45 -POSTHOOK: query: FROM srcpart -INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@dest1_n45 -POSTHOOK: Lineage: dest1_n45.ds SIMPLE [] -POSTHOOK: Lineage: dest1_n45.hr SIMPLE [] -POSTHOOK: Lineage: dest1_n45.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n45.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n45.* FROM dest1_n45 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n45 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n45.* FROM dest1_n45 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n45 -#### A masked pattern was here #### -86 val_86 12 2008-04-08 -27 val_27 12 2008-04-08 -98 val_98 12 2008-04-08 -66 val_66 12 2008-04-08 -37 val_37 12 2008-04-08 -15 val_15 12 2008-04-08 -82 val_82 12 2008-04-08 -17 val_17 12 2008-04-08 -0 val_0 12 2008-04-08 -57 val_57 12 2008-04-08 -20 val_20 12 2008-04-08 -92 val_92 12 2008-04-08 -47 val_47 12 2008-04-08 -72 val_72 12 2008-04-08 -4 val_4 12 2008-04-08 -35 val_35 12 2008-04-08 -54 val_54 12 2008-04-08 -51 val_51 12 2008-04-08 -65 val_65 12 2008-04-08 -83 val_83 12 2008-04-08 -12 val_12 12 2008-04-08 -67 val_67 12 2008-04-08 -84 val_84 12 2008-04-08 -58 val_58 12 2008-04-08 -8 val_8 12 2008-04-08 -24 val_24 12 2008-04-08 -42 val_42 12 2008-04-08 -0 val_0 12 2008-04-08 -96 val_96 12 2008-04-08 -26 val_26 12 2008-04-08 -51 val_51 12 2008-04-08 -43 val_43 12 2008-04-08 -95 val_95 12 2008-04-08 -98 val_98 12 2008-04-08 -85 val_85 12 2008-04-08 -77 val_77 12 2008-04-08 -0 val_0 12 2008-04-08 -87 val_87 12 2008-04-08 -15 val_15 12 2008-04-08 -72 val_72 12 2008-04-08 -90 val_90 12 2008-04-08 -19 val_19 12 2008-04-08 -10 val_10 12 2008-04-08 -5 val_5 12 2008-04-08 -58 val_58 12 2008-04-08 -35 val_35 12 2008-04-08 -95 val_95 12 2008-04-08 -11 val_11 12 2008-04-08 -34 val_34 12 2008-04-08 -42 val_42 12 2008-04-08 -78 val_78 12 2008-04-08 -76 val_76 12 2008-04-08 -41 val_41 12 2008-04-08 -30 val_30 12 2008-04-08 -64 val_64 12 2008-04-08 -76 val_76 12 2008-04-08 -74 val_74 12 2008-04-08 -69 val_69 12 2008-04-08 -33 val_33 12 2008-04-08 -70 val_70 12 2008-04-08 -5 val_5 12 2008-04-08 -2 val_2 12 2008-04-08 -35 val_35 12 2008-04-08 -80 val_80 12 2008-04-08 -44 val_44 12 2008-04-08 -53 val_53 12 2008-04-08 -90 val_90 12 2008-04-08 -12 val_12 12 2008-04-08 -5 val_5 12 2008-04-08 -70 val_70 12 2008-04-08 -24 val_24 12 2008-04-08 -70 val_70 12 2008-04-08 -83 val_83 12 2008-04-08 -26 val_26 12 2008-04-08 -67 val_67 12 2008-04-08 -18 val_18 12 2008-04-08 -9 val_9 12 2008-04-08 -18 val_18 12 2008-04-08 -97 val_97 12 2008-04-08 -84 val_84 12 2008-04-08 -28 val_28 12 2008-04-08 -37 val_37 12 2008-04-08 -90 val_90 12 2008-04-08 -97 val_97 12 2008-04-08 diff --git ql/src/test/results/clientpositive/input_part10.q.out ql/src/test/results/clientpositive/input_part10.q.out deleted file mode 100644 index ead0d14f8e..0000000000 --- ql/src/test/results/clientpositive/input_part10.q.out +++ /dev/null @@ -1,189 +0,0 @@ -PREHOOK: query: CREATE TABLE part_special ( - a STRING, - b STRING -) PARTITIONED BY ( - ds STRING, - ts STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part_special -POSTHOOK: query: CREATE TABLE part_special ( - a STRING, - b STRING -) PARTITIONED BY ( - ds STRING, - ts STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part_special -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -SELECT 1, 2 FROM src LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -SELECT 1, 2 FROM src LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 1 (type: int), 2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: int), _col1 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: CAST( _col0 AS STRING) (type: string), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.part_special - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008 04 08' (type: string), '10:11:12=455' (type: string) - outputColumnNames: a, b, ds, ts - Statistics: Num rows: 1 Data size: 558 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll') - keys: ds (type: string), ts (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008 04 08 - ts 10:11:12=455 - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.part_special - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: a, b - Column Types: string, string - Table: default.part_special - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -SELECT 1, 2 FROM src LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -POSTHOOK: query: INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -SELECT 1, 2 FROM src LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a EXPRESSION [] -POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b EXPRESSION [] -PREHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@part_special -POSTHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@part_special -a string -b string -ds string -ts string - -# Partition Information -# col_name data_type comment -ds string -ts string - -#### A masked pattern was here #### -PREHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' -PREHOOK: type: QUERY -PREHOOK: Input: default@part_special -PREHOOK: Input: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part_special -POSTHOOK: Input: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 -#### A masked pattern was here #### -1 2 2008 04 08 10:11:12=455 diff --git ql/src/test/results/clientpositive/input_part2.q.out ql/src/test/results/clientpositive/input_part2.q.out deleted file mode 100644 index 9d56af795f..0000000000 --- ql/src/test/results/clientpositive/input_part2.q.out +++ /dev/null @@ -1,884 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n84 -POSTHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n84 -PREHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest2_n20 -POSTHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest2_n20 -PREHOOK: query: EXPLAIN EXTENDED -FROM srcpart -INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -PREHOOK: Output: default@dest1_n84 -PREHOOK: Output: default@dest2_n20 -POSTHOOK: query: EXPLAIN EXTENDED -FROM srcpart -INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -POSTHOOK: Output: default@dest1_n84 -POSTHOOK: Output: default@dest2_n20 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 - Stage-3 depends on stages: Stage-0, Stage-10 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-1 depends on stages: Stage-2 - Stage-9 depends on stages: Stage-1, Stage-10 - Stage-10 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - Statistics: Num rows: 1000 Data size: 272000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key < 100) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 333 Data size: 90576 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-08' (type: string) - outputColumnNames: key, value, hr, ds - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Filter Operator - isSamplingPred: false - predicate: ((key < 100) and (ds = '2008-04-09')) (type: boolean) - Statistics: Num rows: 333 Data size: 90576 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-09' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 2 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest2_n20 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n20 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-09' (type: string) - outputColumnNames: key, value, hr, ds - Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-09 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=12 [srcpart] - /srcpart/ds=2008-04-09/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, hr, ds - Column Types: int, string, string, string - Table: default.dest1_n84 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10004 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - name: default.dest1_n84 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10004 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n84 - name: default.dest1_n84 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-1 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,hr,ds - columns.comments - columns.types int:string:string:string -#### A masked pattern was here #### - name default.dest2_n20 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n20 - - Stage: Stage-9 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, hr, ds - Column Types: int, string, string, string - Table: default.dest2_n20 - Is Table Level Stats: true - - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10005 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: FROM srcpart -INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -PREHOOK: Output: default@dest1_n84 -PREHOOK: Output: default@dest2_n20 -POSTHOOK: query: FROM srcpart -INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' -INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -POSTHOOK: Output: default@dest1_n84 -POSTHOOK: Output: default@dest2_n20 -POSTHOOK: Lineage: dest1_n84.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] -POSTHOOK: Lineage: dest1_n84.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] -POSTHOOK: Lineage: dest1_n84.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n84.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n20.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n20.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] -POSTHOOK: Lineage: dest2_n20.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest2_n20.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n84 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n84 -#### A masked pattern was here #### -0 val_0 12 2008-04-08 -0 val_0 12 2008-04-08 -0 val_0 12 2008-04-08 -10 val_10 12 2008-04-08 -11 val_11 12 2008-04-08 -12 val_12 12 2008-04-08 -12 val_12 12 2008-04-08 -15 val_15 12 2008-04-08 -15 val_15 12 2008-04-08 -17 val_17 12 2008-04-08 -18 val_18 12 2008-04-08 -18 val_18 12 2008-04-08 -19 val_19 12 2008-04-08 -2 val_2 12 2008-04-08 -20 val_20 12 2008-04-08 -24 val_24 12 2008-04-08 -24 val_24 12 2008-04-08 -26 val_26 12 2008-04-08 -26 val_26 12 2008-04-08 -27 val_27 12 2008-04-08 -28 val_28 12 2008-04-08 -30 val_30 12 2008-04-08 -33 val_33 12 2008-04-08 -34 val_34 12 2008-04-08 -35 val_35 12 2008-04-08 -35 val_35 12 2008-04-08 -35 val_35 12 2008-04-08 -37 val_37 12 2008-04-08 -37 val_37 12 2008-04-08 -4 val_4 12 2008-04-08 -41 val_41 12 2008-04-08 -42 val_42 12 2008-04-08 -42 val_42 12 2008-04-08 -43 val_43 12 2008-04-08 -44 val_44 12 2008-04-08 -47 val_47 12 2008-04-08 -5 val_5 12 2008-04-08 -5 val_5 12 2008-04-08 -5 val_5 12 2008-04-08 -51 val_51 12 2008-04-08 -51 val_51 12 2008-04-08 -53 val_53 12 2008-04-08 -54 val_54 12 2008-04-08 -57 val_57 12 2008-04-08 -58 val_58 12 2008-04-08 -58 val_58 12 2008-04-08 -64 val_64 12 2008-04-08 -65 val_65 12 2008-04-08 -66 val_66 12 2008-04-08 -67 val_67 12 2008-04-08 -67 val_67 12 2008-04-08 -69 val_69 12 2008-04-08 -70 val_70 12 2008-04-08 -70 val_70 12 2008-04-08 -70 val_70 12 2008-04-08 -72 val_72 12 2008-04-08 -72 val_72 12 2008-04-08 -74 val_74 12 2008-04-08 -76 val_76 12 2008-04-08 -76 val_76 12 2008-04-08 -77 val_77 12 2008-04-08 -78 val_78 12 2008-04-08 -8 val_8 12 2008-04-08 -80 val_80 12 2008-04-08 -82 val_82 12 2008-04-08 -83 val_83 12 2008-04-08 -83 val_83 12 2008-04-08 -84 val_84 12 2008-04-08 -84 val_84 12 2008-04-08 -85 val_85 12 2008-04-08 -86 val_86 12 2008-04-08 -87 val_87 12 2008-04-08 -9 val_9 12 2008-04-08 -90 val_90 12 2008-04-08 -90 val_90 12 2008-04-08 -90 val_90 12 2008-04-08 -92 val_92 12 2008-04-08 -95 val_95 12 2008-04-08 -95 val_95 12 2008-04-08 -96 val_96 12 2008-04-08 -97 val_97 12 2008-04-08 -97 val_97 12 2008-04-08 -98 val_98 12 2008-04-08 -98 val_98 12 2008-04-08 -PREHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr -PREHOOK: type: QUERY -PREHOOK: Input: default@dest2_n20 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest2_n20 -#### A masked pattern was here #### -0 val_0 12 2008-04-09 -0 val_0 12 2008-04-09 -0 val_0 12 2008-04-09 -10 val_10 12 2008-04-09 -11 val_11 12 2008-04-09 -12 val_12 12 2008-04-09 -12 val_12 12 2008-04-09 -15 val_15 12 2008-04-09 -15 val_15 12 2008-04-09 -17 val_17 12 2008-04-09 -18 val_18 12 2008-04-09 -18 val_18 12 2008-04-09 -19 val_19 12 2008-04-09 -2 val_2 12 2008-04-09 -20 val_20 12 2008-04-09 -24 val_24 12 2008-04-09 -24 val_24 12 2008-04-09 -26 val_26 12 2008-04-09 -26 val_26 12 2008-04-09 -27 val_27 12 2008-04-09 -28 val_28 12 2008-04-09 -30 val_30 12 2008-04-09 -33 val_33 12 2008-04-09 -34 val_34 12 2008-04-09 -35 val_35 12 2008-04-09 -35 val_35 12 2008-04-09 -35 val_35 12 2008-04-09 -37 val_37 12 2008-04-09 -37 val_37 12 2008-04-09 -4 val_4 12 2008-04-09 -41 val_41 12 2008-04-09 -42 val_42 12 2008-04-09 -42 val_42 12 2008-04-09 -43 val_43 12 2008-04-09 -44 val_44 12 2008-04-09 -47 val_47 12 2008-04-09 -5 val_5 12 2008-04-09 -5 val_5 12 2008-04-09 -5 val_5 12 2008-04-09 -51 val_51 12 2008-04-09 -51 val_51 12 2008-04-09 -53 val_53 12 2008-04-09 -54 val_54 12 2008-04-09 -57 val_57 12 2008-04-09 -58 val_58 12 2008-04-09 -58 val_58 12 2008-04-09 -64 val_64 12 2008-04-09 -65 val_65 12 2008-04-09 -66 val_66 12 2008-04-09 -67 val_67 12 2008-04-09 -67 val_67 12 2008-04-09 -69 val_69 12 2008-04-09 -70 val_70 12 2008-04-09 -70 val_70 12 2008-04-09 -70 val_70 12 2008-04-09 -72 val_72 12 2008-04-09 -72 val_72 12 2008-04-09 -74 val_74 12 2008-04-09 -76 val_76 12 2008-04-09 -76 val_76 12 2008-04-09 -77 val_77 12 2008-04-09 -78 val_78 12 2008-04-09 -8 val_8 12 2008-04-09 -80 val_80 12 2008-04-09 -82 val_82 12 2008-04-09 -83 val_83 12 2008-04-09 -83 val_83 12 2008-04-09 -84 val_84 12 2008-04-09 -84 val_84 12 2008-04-09 -85 val_85 12 2008-04-09 -86 val_86 12 2008-04-09 -87 val_87 12 2008-04-09 -9 val_9 12 2008-04-09 -90 val_90 12 2008-04-09 -90 val_90 12 2008-04-09 -90 val_90 12 2008-04-09 -92 val_92 12 2008-04-09 -95 val_95 12 2008-04-09 -95 val_95 12 2008-04-09 -96 val_96 12 2008-04-09 -97 val_97 12 2008-04-09 -97 val_97 12 2008-04-09 -98 val_98 12 2008-04-09 -98 val_98 12 2008-04-09 diff --git ql/src/test/results/clientpositive/input_part7.q.out ql/src/test/results/clientpositive/input_part7.q.out deleted file mode 100644 index f040d60180..0000000000 --- ql/src/test/results/clientpositive/input_part7.q.out +++ /dev/null @@ -1,587 +0,0 @@ -PREHOOK: query: EXPLAIN EXTENDED -SELECT * FROM ( - SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 - UNION ALL - SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 -) A -SORT BY A.key, A.value, A.ds, A.hr -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED -SELECT * FROM ( - SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 - UNION ALL - SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 -) A -SORT BY A.key, A.value, A.ds, A.hr -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) - null sort order: zzz - numBuckets: -1 - sort order: +++ - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - auto parallelism: false - TableScan - alias: y - filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) - null sort order: zzz - numBuckets: -1 - sort order: +++ - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0-subquery1:x, $hdt$_0-subquery2:y] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0-subquery1:x, $hdt$_0-subquery2:y] - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), '2008-04-08' (type: string), KEY.reducesinkkey2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM ( - SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 - UNION ALL - SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 -) A -SORT BY A.key, A.value, A.ds, A.hr -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM ( - SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 - UNION ALL - SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 -) A -SORT BY A.key, A.value, A.ds, A.hr -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 11 -0 val_0 2008-04-08 12 -0 val_0 2008-04-08 12 -0 val_0 2008-04-08 12 -0 val_0 2008-04-08 12 -0 val_0 2008-04-08 12 -0 val_0 2008-04-08 12 -10 val_10 2008-04-08 11 -10 val_10 2008-04-08 11 -10 val_10 2008-04-08 12 -10 val_10 2008-04-08 12 -11 val_11 2008-04-08 11 -11 val_11 2008-04-08 11 -11 val_11 2008-04-08 12 -11 val_11 2008-04-08 12 -12 val_12 2008-04-08 11 -12 val_12 2008-04-08 11 -12 val_12 2008-04-08 11 -12 val_12 2008-04-08 11 -12 val_12 2008-04-08 12 -12 val_12 2008-04-08 12 -12 val_12 2008-04-08 12 -12 val_12 2008-04-08 12 -15 val_15 2008-04-08 11 -15 val_15 2008-04-08 11 -15 val_15 2008-04-08 11 -15 val_15 2008-04-08 11 -15 val_15 2008-04-08 12 -15 val_15 2008-04-08 12 -15 val_15 2008-04-08 12 -15 val_15 2008-04-08 12 -17 val_17 2008-04-08 11 -17 val_17 2008-04-08 11 -17 val_17 2008-04-08 12 -17 val_17 2008-04-08 12 -18 val_18 2008-04-08 11 -18 val_18 2008-04-08 11 -18 val_18 2008-04-08 11 -18 val_18 2008-04-08 11 -18 val_18 2008-04-08 12 -18 val_18 2008-04-08 12 -18 val_18 2008-04-08 12 -18 val_18 2008-04-08 12 -19 val_19 2008-04-08 11 -19 val_19 2008-04-08 11 -19 val_19 2008-04-08 12 -19 val_19 2008-04-08 12 -2 val_2 2008-04-08 11 -2 val_2 2008-04-08 11 -2 val_2 2008-04-08 12 -2 val_2 2008-04-08 12 -20 val_20 2008-04-08 11 -20 val_20 2008-04-08 11 -20 val_20 2008-04-08 12 -20 val_20 2008-04-08 12 -24 val_24 2008-04-08 11 -24 val_24 2008-04-08 11 -24 val_24 2008-04-08 11 -24 val_24 2008-04-08 11 -24 val_24 2008-04-08 12 -24 val_24 2008-04-08 12 -24 val_24 2008-04-08 12 -24 val_24 2008-04-08 12 -26 val_26 2008-04-08 11 -26 val_26 2008-04-08 11 -26 val_26 2008-04-08 11 -26 val_26 2008-04-08 11 -26 val_26 2008-04-08 12 -26 val_26 2008-04-08 12 -26 val_26 2008-04-08 12 -26 val_26 2008-04-08 12 -27 val_27 2008-04-08 11 -27 val_27 2008-04-08 11 -27 val_27 2008-04-08 12 -27 val_27 2008-04-08 12 -28 val_28 2008-04-08 11 -28 val_28 2008-04-08 11 -28 val_28 2008-04-08 12 -28 val_28 2008-04-08 12 -30 val_30 2008-04-08 11 -30 val_30 2008-04-08 11 -30 val_30 2008-04-08 12 -30 val_30 2008-04-08 12 -33 val_33 2008-04-08 11 -33 val_33 2008-04-08 11 -33 val_33 2008-04-08 12 -33 val_33 2008-04-08 12 -34 val_34 2008-04-08 11 -34 val_34 2008-04-08 11 -34 val_34 2008-04-08 12 -34 val_34 2008-04-08 12 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 11 -35 val_35 2008-04-08 12 -35 val_35 2008-04-08 12 -35 val_35 2008-04-08 12 -35 val_35 2008-04-08 12 -35 val_35 2008-04-08 12 -35 val_35 2008-04-08 12 -37 val_37 2008-04-08 11 -37 val_37 2008-04-08 11 -37 val_37 2008-04-08 11 -37 val_37 2008-04-08 11 -37 val_37 2008-04-08 12 -37 val_37 2008-04-08 12 -37 val_37 2008-04-08 12 -37 val_37 2008-04-08 12 -4 val_4 2008-04-08 11 -4 val_4 2008-04-08 11 -4 val_4 2008-04-08 12 -4 val_4 2008-04-08 12 -41 val_41 2008-04-08 11 -41 val_41 2008-04-08 11 -41 val_41 2008-04-08 12 -41 val_41 2008-04-08 12 -42 val_42 2008-04-08 11 -42 val_42 2008-04-08 11 -42 val_42 2008-04-08 11 -42 val_42 2008-04-08 11 -42 val_42 2008-04-08 12 -42 val_42 2008-04-08 12 -42 val_42 2008-04-08 12 -42 val_42 2008-04-08 12 -43 val_43 2008-04-08 11 -43 val_43 2008-04-08 11 -43 val_43 2008-04-08 12 -43 val_43 2008-04-08 12 -44 val_44 2008-04-08 11 -44 val_44 2008-04-08 11 -44 val_44 2008-04-08 12 -44 val_44 2008-04-08 12 -47 val_47 2008-04-08 11 -47 val_47 2008-04-08 11 -47 val_47 2008-04-08 12 -47 val_47 2008-04-08 12 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 11 -5 val_5 2008-04-08 12 -5 val_5 2008-04-08 12 -5 val_5 2008-04-08 12 -5 val_5 2008-04-08 12 -5 val_5 2008-04-08 12 -5 val_5 2008-04-08 12 -51 val_51 2008-04-08 11 -51 val_51 2008-04-08 11 -51 val_51 2008-04-08 11 -51 val_51 2008-04-08 11 -51 val_51 2008-04-08 12 -51 val_51 2008-04-08 12 -51 val_51 2008-04-08 12 -51 val_51 2008-04-08 12 -53 val_53 2008-04-08 11 -53 val_53 2008-04-08 11 -53 val_53 2008-04-08 12 -53 val_53 2008-04-08 12 -54 val_54 2008-04-08 11 -54 val_54 2008-04-08 11 -54 val_54 2008-04-08 12 -54 val_54 2008-04-08 12 -57 val_57 2008-04-08 11 -57 val_57 2008-04-08 11 -57 val_57 2008-04-08 12 -57 val_57 2008-04-08 12 -58 val_58 2008-04-08 11 -58 val_58 2008-04-08 11 -58 val_58 2008-04-08 11 -58 val_58 2008-04-08 11 -58 val_58 2008-04-08 12 -58 val_58 2008-04-08 12 -58 val_58 2008-04-08 12 -58 val_58 2008-04-08 12 -64 val_64 2008-04-08 11 -64 val_64 2008-04-08 11 -64 val_64 2008-04-08 12 -64 val_64 2008-04-08 12 -65 val_65 2008-04-08 11 -65 val_65 2008-04-08 11 -65 val_65 2008-04-08 12 -65 val_65 2008-04-08 12 -66 val_66 2008-04-08 11 -66 val_66 2008-04-08 11 -66 val_66 2008-04-08 12 -66 val_66 2008-04-08 12 -67 val_67 2008-04-08 11 -67 val_67 2008-04-08 11 -67 val_67 2008-04-08 11 -67 val_67 2008-04-08 11 -67 val_67 2008-04-08 12 -67 val_67 2008-04-08 12 -67 val_67 2008-04-08 12 -67 val_67 2008-04-08 12 -69 val_69 2008-04-08 11 -69 val_69 2008-04-08 11 -69 val_69 2008-04-08 12 -69 val_69 2008-04-08 12 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 11 -70 val_70 2008-04-08 12 -70 val_70 2008-04-08 12 -70 val_70 2008-04-08 12 -70 val_70 2008-04-08 12 -70 val_70 2008-04-08 12 -70 val_70 2008-04-08 12 -72 val_72 2008-04-08 11 -72 val_72 2008-04-08 11 -72 val_72 2008-04-08 11 -72 val_72 2008-04-08 11 -72 val_72 2008-04-08 12 -72 val_72 2008-04-08 12 -72 val_72 2008-04-08 12 -72 val_72 2008-04-08 12 -74 val_74 2008-04-08 11 -74 val_74 2008-04-08 11 -74 val_74 2008-04-08 12 -74 val_74 2008-04-08 12 -76 val_76 2008-04-08 11 -76 val_76 2008-04-08 11 -76 val_76 2008-04-08 11 -76 val_76 2008-04-08 11 -76 val_76 2008-04-08 12 -76 val_76 2008-04-08 12 -76 val_76 2008-04-08 12 -76 val_76 2008-04-08 12 -77 val_77 2008-04-08 11 -77 val_77 2008-04-08 11 -77 val_77 2008-04-08 12 -77 val_77 2008-04-08 12 -78 val_78 2008-04-08 11 -78 val_78 2008-04-08 11 -78 val_78 2008-04-08 12 -78 val_78 2008-04-08 12 -8 val_8 2008-04-08 11 -8 val_8 2008-04-08 11 -8 val_8 2008-04-08 12 -8 val_8 2008-04-08 12 -80 val_80 2008-04-08 11 -80 val_80 2008-04-08 11 -80 val_80 2008-04-08 12 -80 val_80 2008-04-08 12 -82 val_82 2008-04-08 11 -82 val_82 2008-04-08 11 -82 val_82 2008-04-08 12 -82 val_82 2008-04-08 12 -83 val_83 2008-04-08 11 -83 val_83 2008-04-08 11 -83 val_83 2008-04-08 11 -83 val_83 2008-04-08 11 -83 val_83 2008-04-08 12 -83 val_83 2008-04-08 12 -83 val_83 2008-04-08 12 -83 val_83 2008-04-08 12 -84 val_84 2008-04-08 11 -84 val_84 2008-04-08 11 -84 val_84 2008-04-08 11 -84 val_84 2008-04-08 11 -84 val_84 2008-04-08 12 -84 val_84 2008-04-08 12 -84 val_84 2008-04-08 12 -84 val_84 2008-04-08 12 -85 val_85 2008-04-08 11 -85 val_85 2008-04-08 11 -85 val_85 2008-04-08 12 -85 val_85 2008-04-08 12 -86 val_86 2008-04-08 11 -86 val_86 2008-04-08 11 -86 val_86 2008-04-08 12 -86 val_86 2008-04-08 12 -87 val_87 2008-04-08 11 -87 val_87 2008-04-08 11 -87 val_87 2008-04-08 12 -87 val_87 2008-04-08 12 -9 val_9 2008-04-08 11 -9 val_9 2008-04-08 11 -9 val_9 2008-04-08 12 -9 val_9 2008-04-08 12 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 11 -90 val_90 2008-04-08 12 -90 val_90 2008-04-08 12 -90 val_90 2008-04-08 12 -90 val_90 2008-04-08 12 -90 val_90 2008-04-08 12 -90 val_90 2008-04-08 12 -92 val_92 2008-04-08 11 -92 val_92 2008-04-08 11 -92 val_92 2008-04-08 12 -92 val_92 2008-04-08 12 -95 val_95 2008-04-08 11 -95 val_95 2008-04-08 11 -95 val_95 2008-04-08 11 -95 val_95 2008-04-08 11 -95 val_95 2008-04-08 12 -95 val_95 2008-04-08 12 -95 val_95 2008-04-08 12 -95 val_95 2008-04-08 12 -96 val_96 2008-04-08 11 -96 val_96 2008-04-08 11 -96 val_96 2008-04-08 12 -96 val_96 2008-04-08 12 -97 val_97 2008-04-08 11 -97 val_97 2008-04-08 11 -97 val_97 2008-04-08 11 -97 val_97 2008-04-08 11 -97 val_97 2008-04-08 12 -97 val_97 2008-04-08 12 -97 val_97 2008-04-08 12 -97 val_97 2008-04-08 12 -98 val_98 2008-04-08 11 -98 val_98 2008-04-08 11 -98 val_98 2008-04-08 11 -98 val_98 2008-04-08 11 -98 val_98 2008-04-08 12 -98 val_98 2008-04-08 12 -98 val_98 2008-04-08 12 -98 val_98 2008-04-08 12 diff --git ql/src/test/results/clientpositive/input_testxpath.q.out ql/src/test/results/clientpositive/input_testxpath.q.out deleted file mode 100644 index c400c0d4b9..0000000000 --- ql/src/test/results/clientpositive/input_testxpath.q.out +++ /dev/null @@ -1,167 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n6(key INT, value STRING, mapvalue STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n6 -POSTHOOK: query: CREATE TABLE dest1_n6(key INT, value STRING, mapvalue STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n6 -PREHOOK: query: EXPLAIN -FROM src_thrift -INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -PREHOOK: Output: default@dest1_n6 -POSTHOOK: query: EXPLAIN -FROM src_thrift -INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -POSTHOOK: Output: default@dest1_n6 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: lint[1] (type: int), lintstring[0].mystring (type: string), mstringstring['key_2'] (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n6 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, mapvalue - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(mapvalue, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 4904 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 4904 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n6 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value, mapvalue - Column Types: int, string, string - Table: default.dest1_n6 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n6 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n6 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src_thrift -INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -PREHOOK: Output: default@dest1_n6 -POSTHOOK: query: FROM src_thrift -INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -POSTHOOK: Output: default@dest1_n6 -POSTHOOK: Lineage: dest1_n6.key EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lint, type:array, comment:from deserializer), ] -POSTHOOK: Lineage: dest1_n6.mapvalue EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:mstringstring, type:map, comment:from deserializer), ] -POSTHOOK: Lineage: dest1_n6.value EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lintstring, type:array>, comment:from deserializer), ] -PREHOOK: query: SELECT dest1_n6.* FROM dest1_n6 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n6 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n6.* FROM dest1_n6 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n6 -#### A masked pattern was here #### -0 0 NULL -2 1 NULL -4 8 value_2 -6 27 NULL -8 64 NULL -10 125 NULL -12 216 NULL -14 343 NULL -16 512 NULL -18 729 NULL -NULL NULL NULL diff --git ql/src/test/results/clientpositive/input_testxpath2.q.out ql/src/test/results/clientpositive/input_testxpath2.q.out deleted file mode 100644 index 408bc6a504..0000000000 --- ql/src/test/results/clientpositive/input_testxpath2.q.out +++ /dev/null @@ -1,170 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n32(lint_size INT, lintstring_size INT, mstringstring_size INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n32 -POSTHOOK: query: CREATE TABLE dest1_n32(lint_size INT, lintstring_size INT, mstringstring_size INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n32 -PREHOOK: query: EXPLAIN -FROM src_thrift -INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -PREHOOK: Output: default@dest1_n32 -POSTHOOK: query: EXPLAIN -FROM src_thrift -INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -POSTHOOK: Output: default@dest1_n32 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - filterExpr: (lint is not null and mstringstring is not null) (type: boolean) - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (lint is not null and mstringstring is not null) (type: boolean) - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: size(lint) (type: int), size(lintstring) (type: int), size(mstringstring) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n32 - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: lint_size, lintstring_size, mstringstring_size - Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(lint_size, 'hll'), compute_stats(lintstring_size, 'hll'), compute_stats(mstringstring_size, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 4872 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 4872 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n32 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: lint_size, lintstring_size, mstringstring_size - Column Types: int, int, int - Table: default.dest1_n32 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n32 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n32 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: FROM src_thrift -INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -PREHOOK: Output: default@dest1_n32 -POSTHOOK: query: FROM src_thrift -INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -POSTHOOK: Output: default@dest1_n32 -POSTHOOK: Lineage: dest1_n32.lint_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lint, type:array, comment:from deserializer), ] -POSTHOOK: Lineage: dest1_n32.lintstring_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lintstring, type:array>, comment:from deserializer), ] -POSTHOOK: Lineage: dest1_n32.mstringstring_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:mstringstring, type:map, comment:from deserializer), ] -PREHOOK: query: SELECT dest1_n32.* FROM dest1_n32 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n32 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n32.* FROM dest1_n32 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n32 -#### A masked pattern was here #### -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 -3 1 1 diff --git ql/src/test/results/clientpositive/insert2_overwrite_partitions.q.out ql/src/test/results/clientpositive/insert2_overwrite_partitions.q.out deleted file mode 100644 index 8d571b68f5..0000000000 --- ql/src/test/results/clientpositive/insert2_overwrite_partitions.q.out +++ /dev/null @@ -1,379 +0,0 @@ -PREHOOK: query: CREATE DATABASE db1 -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:db1 -POSTHOOK: query: CREATE DATABASE db1 -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:db1 -PREHOOK: query: CREATE DATABASE db2 -PREHOOK: type: CREATEDATABASE -PREHOOK: Output: database:db2 -POSTHOOK: query: CREATE DATABASE db2 -POSTHOOK: type: CREATEDATABASE -POSTHOOK: Output: database:db2 -PREHOOK: query: CREATE TABLE db1.sourceTable (one string,two string) PARTITIONED BY (ds string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:db1 -PREHOOK: Output: db1@sourceTable -POSTHOOK: query: CREATE TABLE db1.sourceTable (one string,two string) PARTITIONED BY (ds string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:db1 -POSTHOOK: Output: db1@sourceTable -PREHOOK: query: load data local inpath '../../data/files/kv1.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: db1@sourcetable -POSTHOOK: query: load data local inpath '../../data/files/kv1.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: db1@sourcetable -POSTHOOK: Output: db1@sourcetable@ds=2011-11-11 -PREHOOK: query: load data local inpath '../../data/files/kv3.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: db1@sourcetable@ds=2011-11-11 -POSTHOOK: query: load data local inpath '../../data/files/kv3.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: db1@sourcetable@ds=2011-11-11 -PREHOOK: query: CREATE TABLE db2.destinTable (one string,two string) PARTITIONED BY (ds string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:db2 -PREHOOK: Output: db2@destinTable -POSTHOOK: query: CREATE TABLE db2.destinTable (one string,two string) PARTITIONED BY (ds string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:db2 -POSTHOOK: Output: db2@destinTable -PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: db1@sourcetable -PREHOOK: Input: db1@sourcetable@ds=2011-11-11 -PREHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: db1@sourcetable -POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 -POSTHOOK: Output: db2@destintable@ds=2011-11-11 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: sourcetable - filterExpr: (ds = '2011-11-11') (type: boolean) - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: one (type: string), two (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -- - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: db2.destintable - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2011-11-11' (type: string) - outputColumnNames: one, two, ds - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: compute_stats(one, 'hll'), compute_stats(two, 'hll') - keys: ds (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2011-11-11 - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: db2.destintable - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: one, two - Column Types: string, string - Table: db2.destintable - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - value expressions: _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: db1@sourcetable -PREHOOK: Input: db1@sourcetable@ds=2011-11-11 -PREHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: db1@sourcetable -POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 -POSTHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).one SIMPLE [(sourcetable)sourcetable.FieldSchema(name:one, type:string, comment:null), ] -POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).two SIMPLE [(sourcetable)sourcetable.FieldSchema(name:two, type:string, comment:null), ] -PREHOOK: query: select one,two from db2.destinTable order by one desc, two desc -PREHOOK: type: QUERY -PREHOOK: Input: db2@destintable -PREHOOK: Input: db2@destintable@ds=2011-11-11 -#### A masked pattern was here #### -POSTHOOK: query: select one,two from db2.destinTable order by one desc, two desc -POSTHOOK: type: QUERY -POSTHOOK: Input: db2@destintable -POSTHOOK: Input: db2@destintable@ds=2011-11-11 -#### A masked pattern was here #### -98 val_98 -98 val_98 -98 val_98 -97 val_97 -97 val_97 -PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: db1@sourcetable -PREHOOK: Input: db1@sourcetable@ds=2011-11-11 -PREHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: db1@sourcetable -POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 -POSTHOOK: Output: db2@destintable@ds=2011-11-11 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: sourcetable - filterExpr: (ds = '2011-11-11') (type: boolean) - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: one (type: string), two (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -- - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE - Limit - Number of rows: 5 - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: db2.destintable - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2011-11-11' (type: string) - outputColumnNames: one, two, ds - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: compute_stats(one, 'hll'), compute_stats(two, 'hll') - keys: ds (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2011-11-11 - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: db2.destintable - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: one, two - Column Types: string, string - Table: db2.destintable - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE - value expressions: _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -PREHOOK: type: QUERY -PREHOOK: Input: db1@sourcetable -PREHOOK: Input: db1@sourcetable@ds=2011-11-11 -PREHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') -SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 -POSTHOOK: type: QUERY -POSTHOOK: Input: db1@sourcetable -POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 -POSTHOOK: Output: db2@destintable@ds=2011-11-11 -POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).one SIMPLE [(sourcetable)sourcetable.FieldSchema(name:one, type:string, comment:null), ] -POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).two SIMPLE [(sourcetable)sourcetable.FieldSchema(name:two, type:string, comment:null), ] -PREHOOK: query: select one,two from db2.destinTable order by one desc, two desc -PREHOOK: type: QUERY -PREHOOK: Input: db2@destintable -PREHOOK: Input: db2@destintable@ds=2011-11-11 -#### A masked pattern was here #### -POSTHOOK: query: select one,two from db2.destinTable order by one desc, two desc -POSTHOOK: type: QUERY -POSTHOOK: Input: db2@destintable -POSTHOOK: Input: db2@destintable@ds=2011-11-11 -#### A masked pattern was here #### -98 val_98 -98 val_98 -98 val_98 -97 val_97 -97 val_97 -PREHOOK: query: drop table db2.destinTable -PREHOOK: type: DROPTABLE -PREHOOK: Input: db2@destintable -PREHOOK: Output: db2@destintable -POSTHOOK: query: drop table db2.destinTable -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: db2@destintable -POSTHOOK: Output: db2@destintable -PREHOOK: query: drop table db1.sourceTable -PREHOOK: type: DROPTABLE -PREHOOK: Input: db1@sourcetable -PREHOOK: Output: db1@sourcetable -POSTHOOK: query: drop table db1.sourceTable -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: db1@sourcetable -POSTHOOK: Output: db1@sourcetable -PREHOOK: query: DROP DATABASE db1 -PREHOOK: type: DROPDATABASE -PREHOOK: Input: database:db1 -PREHOOK: Output: database:db1 -POSTHOOK: query: DROP DATABASE db1 -POSTHOOK: type: DROPDATABASE -POSTHOOK: Input: database:db1 -POSTHOOK: Output: database:db1 -PREHOOK: query: DROP DATABASE db2 -PREHOOK: type: DROPDATABASE -PREHOOK: Input: database:db2 -PREHOOK: Output: database:db2 -POSTHOOK: query: DROP DATABASE db2 -POSTHOOK: type: DROPDATABASE -POSTHOOK: Input: database:db2 -POSTHOOK: Output: database:db2 diff --git ql/src/test/results/clientpositive/insertoverwrite_bucket.q.out ql/src/test/results/clientpositive/insertoverwrite_bucket.q.out deleted file mode 100644 index c02cb2dae9..0000000000 --- ql/src/test/results/clientpositive/insertoverwrite_bucket.q.out +++ /dev/null @@ -1,373 +0,0 @@ -PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketinput( -data string -) -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucketinput -POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketinput( -data string -) -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucketinput -PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput1( -data string -)CLUSTERED BY(data) -INTO 2 BUCKETS -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucketoutput1 -POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput1( -data string -)CLUSTERED BY(data) -INTO 2 BUCKETS -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucketoutput1 -PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput2( -data string -)CLUSTERED BY(data) -INTO 2 BUCKETS -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bucketoutput2 -POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput2( -data string -)CLUSTERED BY(data) -INTO 2 BUCKETS -ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bucketoutput2 -PREHOOK: query: insert into table bucketinput values ("firstinsert1") -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@bucketinput -POSTHOOK: query: insert into table bucketinput values ("firstinsert1") -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@bucketinput -POSTHOOK: Lineage: bucketinput.data SCRIPT [] -PREHOOK: query: insert into table bucketinput values ("firstinsert2") -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@bucketinput -POSTHOOK: query: insert into table bucketinput values ("firstinsert2") -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@bucketinput -POSTHOOK: Lineage: bucketinput.data SCRIPT [] -PREHOOK: query: insert into table bucketinput values ("firstinsert3") -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@bucketinput -POSTHOOK: query: insert into table bucketinput values ("firstinsert3") -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@bucketinput -POSTHOOK: Lineage: bucketinput.data SCRIPT [] -PREHOOK: query: insert overwrite table bucketoutput1 select * from bucketinput where data like 'first%' -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketinput -PREHOOK: Output: default@bucketoutput1 -POSTHOOK: query: insert overwrite table bucketoutput1 select * from bucketinput where data like 'first%' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketinput -POSTHOOK: Output: default@bucketoutput1 -POSTHOOK: Lineage: bucketoutput1.data SIMPLE [(bucketinput)bucketinput.FieldSchema(name:data, type:string, comment:null), ] -PREHOOK: query: CREATE TABLE temp1 -( - change string, - num string -) -CLUSTERED BY (num) SORTED BY (num) INTO 4 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@temp1 -POSTHOOK: query: CREATE TABLE temp1 -( - change string, - num string -) -CLUSTERED BY (num) SORTED BY (num) INTO 4 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@temp1 -PREHOOK: query: explain insert overwrite table temp1 select data, data from bucketinput -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketinput -PREHOOK: Output: default@temp1 -POSTHOOK: query: explain insert overwrite table temp1 select data, data from bucketinput -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketinput -POSTHOOK: Output: default@temp1 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: bucketinput - Statistics: Num rows: 3 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: data (type: string) - outputColumnNames: _col1 - Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: a - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.temp1 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: change, num - Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(change, 'hll'), compute_stats(num, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.temp1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: change, num - Column Types: string, string - Table: default.temp1 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: CREATE TABLE temp2 -( - create_ts STRING , - change STRING, - num STRING -) -CLUSTERED BY (create_ts) SORTED BY (num) INTO 4 BUCKETS -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@temp2 -POSTHOOK: query: CREATE TABLE temp2 -( - create_ts STRING , - change STRING, - num STRING -) -CLUSTERED BY (create_ts) SORTED BY (num) INTO 4 BUCKETS -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@temp2 -PREHOOK: query: explain -INSERT OVERWRITE TABLE temp2 -SELECT change, change,num -FROM temp1 -PREHOOK: type: QUERY -PREHOOK: Input: default@temp1 -PREHOOK: Output: default@temp2 -POSTHOOK: query: explain -INSERT OVERWRITE TABLE temp2 -SELECT change, change,num -FROM temp1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@temp1 -POSTHOOK: Output: default@temp2 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: temp1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: change (type: string), num (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: a - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.temp2 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: create_ts, change, num - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(create_ts, 'hll'), compute_stats(change, 'hll'), compute_stats(num, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.temp2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: create_ts, change, num - Column Types: string, string, string - Table: default.temp2 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: select * from bucketoutput1 a join bucketoutput2 b on (a.data=b.data) -PREHOOK: type: QUERY -PREHOOK: Input: default@bucketoutput1 -PREHOOK: Input: default@bucketoutput2 -#### A masked pattern was here #### -POSTHOOK: query: select * from bucketoutput1 a join bucketoutput2 b on (a.data=b.data) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bucketoutput1 -POSTHOOK: Input: default@bucketoutput2 -#### A masked pattern was here #### -PREHOOK: query: drop table temp1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@temp1 -PREHOOK: Output: default@temp1 -POSTHOOK: query: drop table temp1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@temp1 -POSTHOOK: Output: default@temp1 -PREHOOK: query: drop table temp2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@temp2 -PREHOOK: Output: default@temp2 -POSTHOOK: query: drop table temp2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@temp2 -POSTHOOK: Output: default@temp2 -PREHOOK: query: drop table buckettestinput -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table buckettestinput -POSTHOOK: type: DROPTABLE -PREHOOK: query: drop table buckettestoutput1 -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table buckettestoutput1 -POSTHOOK: type: DROPTABLE -PREHOOK: query: drop table buckettestoutput2 -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table buckettestoutput2 -POSTHOOK: type: DROPTABLE diff --git ql/src/test/results/clientpositive/join11.q.out ql/src/test/results/clientpositive/join11.q.out deleted file mode 100644 index de2e420bb6..0000000000 --- ql/src/test/results/clientpositive/join11.q.out +++ /dev/null @@ -1,256 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -19 val_19 -2 val_2 -20 val_20 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -4 val_4 -41 val_41 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -8 val_8 -80 val_80 -82 val_82 -83 val_83 -83 val_83 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -87 val_87 -9 val_9 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -95 val_95 -95 val_95 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -97 val_97 -97 val_97 -98 val_98 -98 val_98 -98 val_98 -98 val_98 diff --git ql/src/test/results/clientpositive/join12.q.out ql/src/test/results/clientpositive/join12.q.out deleted file mode 100644 index cb668295fe..0000000000 --- ql/src/test/results/clientpositive/join12.q.out +++ /dev/null @@ -1,396 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 = src3.c5 AND src3.c5 < 80 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 = src3.c5 AND src3.c5 < 80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 80.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 = src3.c5 AND src3.c5 < 80 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 = src3.c5 AND src3.c5 < 80 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -18 val_18 -19 val_19 -2 val_2 -20 val_20 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -4 val_4 -41 val_41 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -8 val_8 -9 val_9 diff --git ql/src/test/results/clientpositive/join13.q.out ql/src/test/results/clientpositive/join13.q.out deleted file mode 100644 index 821772be84..0000000000 --- ql/src/test/results/clientpositive/join13.q.out +++ /dev/null @@ -1,329 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 166 Data size: 43990 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) - Statistics: Num rows: 166 Data size: 43990 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string) - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 200.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 200.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToDouble(key) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 166 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) - 1 _col0 (type: double) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT src1.c1, src2.c4 -FROM -(SELECT src.key as c1, src.value as c2 from src) src1 -JOIN -(SELECT src.key as c3, src.value as c4 from src) src2 -ON src1.c1 = src2.c3 AND src1.c1 < 100 -JOIN -(SELECT src.key as c5, src.value as c6 from src) src3 -ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -0 val_0 -10 val_10 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -15 val_15 -15 val_15 -17 val_17 -2 val_2 -27 val_27 -33 val_33 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -37 val_37 -37 val_37 -4 val_4 -41 val_41 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -42 val_42 -43 val_43 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -5 val_5 -57 val_57 -58 val_58 -58 val_58 -58 val_58 -58 val_58 -64 val_64 -64 val_64 -64 val_64 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -67 val_67 -69 val_69 -69 val_69 -69 val_69 -69 val_69 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -76 val_76 -78 val_78 -80 val_80 -82 val_82 -82 val_82 -83 val_83 -83 val_83 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -86 val_86 -87 val_87 -87 val_87 -9 val_9 -9 val_9 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -90 val_90 -95 val_95 -95 val_95 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -97 val_97 -97 val_97 -98 val_98 -98 val_98 -98 val_98 -98 val_98 diff --git ql/src/test/results/clientpositive/join16.q.out ql/src/test/results/clientpositive/join16.q.out deleted file mode 100644 index 27633dfcad..0000000000 --- ql/src/test/results/clientpositive/join16.q.out +++ /dev/null @@ -1,77 +0,0 @@ -PREHOOK: query: EXPLAIN SELECT subq.key, tab.value FROM (select a.key, a.value from src a where a.key > 10 ) subq JOIN src tab ON (subq.key = tab.key and subq.key > 20 and subq.value = tab.value) where tab.value < 200 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT subq.key, tab.value FROM (select a.key, a.value from src a where a.key > 10 ) subq JOIN src tab ON (subq.key = tab.key and subq.key > 20 and subq.value = tab.value) where tab.value < 200 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: tab - filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col1 (type: string) - 1 _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join19.q.out ql/src/test/results/clientpositive/join19.q.out deleted file mode 100644 index 9de353b2d5..0000000000 --- ql/src/test/results/clientpositive/join19.q.out +++ /dev/null @@ -1,377 +0,0 @@ -PREHOOK: query: CREATE TABLE triples (foo string, subject string, predicate string, object string, foo2 string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@triples -POSTHOOK: query: CREATE TABLE triples (foo string, subject string, predicate string, object string, foo2 string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@triples -PREHOOK: query: EXPLAIN -SELECT t11.subject, t22.object , t33.subject , t55.object, t66.object -FROM -( -SELECT t1.subject -FROM triples t1 -WHERE -t1.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' -AND -t1.object='http://ontos/OntosMiner/Common.English/ontology#Citation' -) t11 -JOIN -( -SELECT t2.subject , t2.object -FROM triples t2 -WHERE -t2.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' -) t22 -ON (t11.subject=t22.subject) -JOIN -( -SELECT t3.subject , t3.object -FROM triples t3 -WHERE -t3.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_from' - -) t33 -ON (t11.subject=t33.object) -JOIN -( -SELECT t4.subject -FROM triples t4 -WHERE -t4.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' -AND -t4.object='http://ontos/OntosMiner/Common.English/ontology#Author' - -) t44 -ON (t44.subject=t33.subject) -JOIN -( -SELECT t5.subject, t5.object -FROM triples t5 -WHERE -t5.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_to' -) t55 -ON (t55.subject=t44.subject) -JOIN -( -SELECT t6.subject, t6.object -FROM triples t6 -WHERE -t6.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' -) t66 -ON (t66.subject=t55.object) -PREHOOK: type: QUERY -PREHOOK: Input: default@triples -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT t11.subject, t22.object , t33.subject , t55.object, t66.object -FROM -( -SELECT t1.subject -FROM triples t1 -WHERE -t1.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' -AND -t1.object='http://ontos/OntosMiner/Common.English/ontology#Citation' -) t11 -JOIN -( -SELECT t2.subject , t2.object -FROM triples t2 -WHERE -t2.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' -) t22 -ON (t11.subject=t22.subject) -JOIN -( -SELECT t3.subject , t3.object -FROM triples t3 -WHERE -t3.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_from' - -) t33 -ON (t11.subject=t33.object) -JOIN -( -SELECT t4.subject -FROM triples t4 -WHERE -t4.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' -AND -t4.object='http://ontos/OntosMiner/Common.English/ontology#Author' - -) t44 -ON (t44.subject=t33.subject) -JOIN -( -SELECT t5.subject, t5.object -FROM triples t5 -WHERE -t5.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_to' -) t55 -ON (t55.subject=t44.subject) -JOIN -( -SELECT t6.subject, t6.object -FROM triples t6 -WHERE -t6.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' -) t66 -ON (t66.subject=t55.object) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@triples -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-5 - Stage-3 depends on stages: Stage-2, Stage-7 - Stage-5 is a root stage - Stage-7 is a root stage - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t5 - filterExpr: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') and subject is not null and object is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') and subject is not null and object is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string), object (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string) - TableScan - alias: t6 - filterExpr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string), object (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col3 (type: string) - TableScan - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col2 (type: string) - outputColumnNames: _col1, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 667 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col5 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col5 (type: string) - Statistics: Num rows: 1 Data size: 667 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col3 (type: string), _col4 (type: string) - TableScan - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col5 (type: string) - 1 _col2 (type: string) - outputColumnNames: _col1, _col3, _col4, _col8, _col9 - Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col9 (type: string), _col8 (type: string), _col4 (type: string), _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: t3 - filterExpr: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') and object is not null and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') and object is not null and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string), object (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: t4 - filterExpr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - alias: t2 - filterExpr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string), object (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: t1 - filterExpr: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation') and subject is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: subject (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join22.q.out ql/src/test/results/clientpositive/join22.q.out deleted file mode 100644 index 109779db50..0000000000 --- ql/src/test/results/clientpositive/join22.q.out +++ /dev/null @@ -1,125 +0,0 @@ -PREHOOK: query: explain -SELECT src5.src1_value FROM (SELECT src3.*, src4.value as src4_value, src4.key as src4_key FROM src src4 JOIN (SELECT src2.*, src1.key as src1_key, src1.value as src1_value FROM src src1 JOIN src src2 ON src1.key = src2.key) src3 ON src3.src1_key = src4.key) src5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT src5.src1_value FROM (SELECT src3.*, src4.value as src4_value, src4.key as src4_key FROM src src4 JOIN (SELECT src2.*, src1.key as src1_key, src1.value as src1_value FROM src src1 JOIN src src2 ON src1.key = src2.key) src3 ON src3.src1_key = src4.key) src5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src4 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1 - Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join23.q.out ql/src/test/results/clientpositive/join23.q.out deleted file mode 100644 index 23b4207b4e..0000000000 --- ql/src/test/results/clientpositive/join23.q.out +++ /dev/null @@ -1,206 +0,0 @@ -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - null sort order: zzzz - sort order: ++++ - Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 0 val_0 -0 val_0 2 val_2 -0 val_0 2 val_2 -0 val_0 2 val_2 -0 val_0 4 val_4 -0 val_0 4 val_4 -0 val_0 4 val_4 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 5 val_5 -0 val_0 8 val_8 -0 val_0 8 val_8 -0 val_0 8 val_8 -0 val_0 9 val_9 -0 val_0 9 val_9 -0 val_0 9 val_9 -2 val_2 0 val_0 -2 val_2 0 val_0 -2 val_2 0 val_0 -2 val_2 2 val_2 -2 val_2 4 val_4 -2 val_2 5 val_5 -2 val_2 5 val_5 -2 val_2 5 val_5 -2 val_2 8 val_8 -2 val_2 9 val_9 -4 val_4 0 val_0 -4 val_4 0 val_0 -4 val_4 0 val_0 -4 val_4 2 val_2 -4 val_4 4 val_4 -4 val_4 5 val_5 -4 val_4 5 val_5 -4 val_4 5 val_5 -4 val_4 8 val_8 -4 val_4 9 val_9 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 0 val_0 -5 val_5 2 val_2 -5 val_5 2 val_2 -5 val_5 2 val_2 -5 val_5 4 val_4 -5 val_5 4 val_4 -5 val_5 4 val_4 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 5 val_5 -5 val_5 8 val_8 -5 val_5 8 val_8 -5 val_5 8 val_8 -5 val_5 9 val_9 -5 val_5 9 val_9 -5 val_5 9 val_9 -8 val_8 0 val_0 -8 val_8 0 val_0 -8 val_8 0 val_0 -8 val_8 2 val_2 -8 val_8 4 val_4 -8 val_8 5 val_5 -8 val_8 5 val_5 -8 val_8 5 val_5 -8 val_8 8 val_8 -8 val_8 9 val_9 -9 val_9 0 val_0 -9 val_9 0 val_0 -9 val_9 0 val_0 -9 val_9 2 val_2 -9 val_9 4 val_4 -9 val_9 5 val_5 -9 val_9 5 val_5 -9 val_9 5 val_5 -9 val_9 8 val_8 -9 val_9 9 val_9 diff --git ql/src/test/results/clientpositive/join25.q.out ql/src/test/results/clientpositive/join25.q.out deleted file mode 100644 index b3a613cb1d..0000000000 --- ql/src/test/results/clientpositive/join25.q.out +++ /dev/null @@ -1,213 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n18(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n18 -POSTHOOK: query: CREATE TABLE dest_j1_n18(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n18 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n18 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n18 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n18 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n18 -STAGE DEPENDENCIES: - Stage-6 is a root stage - Stage-5 depends on stages: Stage-6 - Stage-0 depends on stages: Stage-5 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-6 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n18 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n18 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value, val2 - Column Types: int, string, string - Table: default.dest_j1_n18 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n18 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n18 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n18 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n18 -POSTHOOK: Lineage: dest_j1_n18.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n18.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n18.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n18 x -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n18 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n18 x -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n18 -#### A masked pattern was here #### -128 val_128 -128 val_128 -128 val_128 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -224 val_224 -224 val_224 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -369 val_369 -369 val_369 -369 val_369 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 diff --git ql/src/test/results/clientpositive/join26.q.out ql/src/test/results/clientpositive/join26.q.out deleted file mode 100644 index 2031b58b11..0000000000 --- ql/src/test/results/clientpositive/join26.q.out +++ /dev/null @@ -1,651 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n10(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n10 -POSTHOOK: query: CREATE TABLE dest_j1_n10(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n10 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n10 -SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n10 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n10 -SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n10 -OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` -FROM (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `key` IS NOT NULL) AS `t0` -INNER JOIN ((SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` IS NOT NULL) AS `t2` -INNER JOIN (SELECT `key` -FROM `default`.`src1` -WHERE `key` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`key` = `t4`.`key` -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-7 depends on stages: Stage-9 - Stage-0 depends on stages: Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:z - Fetch Operator - limit: -1 - Partition Description: - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - $hdt$_2:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:z - TableScan - alias: z - filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D) and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col2 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - $hdt$_2:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2 - Position of Big Table: 0 - Statistics: Num rows: 39 Data size: 6903 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col4 - Position of Big Table: 0 - Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n10 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n10 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n10 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Local Work: - Map Reduce Local Work - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: src1 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src1 - name: default.src1 -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:y] - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n10 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n10 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n10 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, val2 - Column Types: string, string, string - Table: default.dest_j1_n10 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n10 -SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n10 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n10 -SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n10 -POSTHOOK: Lineage: dest_j1_n10.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n10.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n10.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n10 x -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n10 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n10 x -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n10 -#### A masked pattern was here #### -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -128 val_128 val_128 -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -224 val_224 val_224 -224 val_224 val_224 -224 val_224 val_224 -224 val_224 val_224 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -369 val_369 val_369 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 diff --git ql/src/test/results/clientpositive/join27.q.out ql/src/test/results/clientpositive/join27.q.out deleted file mode 100644 index e1d7b0e6e5..0000000000 --- ql/src/test/results/clientpositive/join27.q.out +++ /dev/null @@ -1,217 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n2(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n2 -POSTHOOK: query: CREATE TABLE dest_j1_n2(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n2 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n2 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.value = y.value) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n2 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n2 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.value = y.value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n2 -STAGE DEPENDENCIES: - Stage-6 is a root stage - Stage-5 depends on stages: Stage-6 - Stage-0 depends on stages: Stage-5 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-6 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:x - TableScan - alias: x - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 40 Data size: 10640 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n2 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value, val2 - Column Types: int, string, string - Table: default.dest_j1_n2 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n2 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.value = y.value) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n2 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n2 -SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.value = y.value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n2 -POSTHOOK: Lineage: dest_j1_n2.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n2.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n2.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n2 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n2 -#### A masked pattern was here #### -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 -NULL val_165 val_165 -NULL val_165 val_165 -NULL val_193 val_193 -NULL val_193 val_193 -NULL val_193 val_193 -NULL val_265 val_265 -NULL val_265 val_265 -NULL val_27 val_27 -NULL val_409 val_409 -NULL val_409 val_409 -NULL val_409 val_409 -NULL val_484 val_484 diff --git ql/src/test/results/clientpositive/join28.q.out ql/src/test/results/clientpositive/join28.q.out deleted file mode 100644 index b1a559ceea..0000000000 --- ql/src/test/results/clientpositive/join28.q.out +++ /dev/null @@ -1,329 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n11(key STRING, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n11 -POSTHOOK: query: CREATE TABLE dest_j1_n11(key STRING, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n11 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n11 -SELECT subq.key1, z.value -FROM -(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 - FROM src1 x JOIN src y ON (x.key = y.key)) subq - JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n11 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n11 -SELECT subq.key1, z.value -FROM -(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 - FROM src1 x JOIN src y ON (x.key = y.key)) subq - JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n11 -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-7 depends on stages: Stage-9 - Stage-0 depends on stages: Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:z - Fetch Operator - limit: -1 - $hdt$_2:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:z - TableScan - alias: z - filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D) and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - $hdt$_2:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1 - Statistics: Num rows: 39 Data size: 3354 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n11 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n11 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.dest_j1_n11 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n11 -SELECT subq.key1, z.value -FROM -(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 - FROM src1 x JOIN src y ON (x.key = y.key)) subq - JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n11 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n11 -SELECT subq.key1, z.value -FROM -(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 - FROM src1 x JOIN src y ON (x.key = y.key)) subq - JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n11 -POSTHOOK: Lineage: dest_j1_n11.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n11.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n11 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n11 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n11 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n11 -#### A masked pattern was here #### -128 val_128 -128 val_128 -128 val_128 -128 val_128 -128 val_128 -128 val_128 -128 val_128 -128 val_128 -128 val_128 -146 val_146 -146 val_146 -146 val_146 -146 val_146 -150 val_150 -213 val_213 -213 val_213 -213 val_213 -213 val_213 -224 val_224 -224 val_224 -224 val_224 -224 val_224 -238 val_238 -238 val_238 -238 val_238 -238 val_238 -255 val_255 -255 val_255 -255 val_255 -255 val_255 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -273 val_273 -278 val_278 -278 val_278 -278 val_278 -278 val_278 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -311 val_311 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -369 val_369 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -401 val_401 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -406 val_406 -66 val_66 -98 val_98 -98 val_98 -98 val_98 -98 val_98 diff --git ql/src/test/results/clientpositive/join29.q.out ql/src/test/results/clientpositive/join29.q.out deleted file mode 100644 index eebc36a497..0000000000 --- ql/src/test/results/clientpositive/join29.q.out +++ /dev/null @@ -1,376 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n6(key STRING, cnt1 INT, cnt2 INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n6 -POSTHOOK: query: CREATE TABLE dest_j1_n6(key STRING, cnt1 INT, cnt2 INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n6 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n6 -SELECT subq1.key, subq1.cnt, subq2.cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n6 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n6 -SELECT subq1.key, subq1.cnt, subq2.cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n6 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-8 depends on stages: Stage-1, Stage-5 , consists of Stage-9, Stage-10, Stage-2 - Stage-9 has a backup stage: Stage-2 - Stage-6 depends on stages: Stage-9 - Stage-0 depends on stages: Stage-2, Stage-6, Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2, Stage-6, Stage-7 - Stage-10 has a backup stage: Stage-2 - Stage-7 depends on stages: Stage-10 - Stage-2 - Stage-5 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-9 - Map Reduce Local Work - Alias -> Map Local Tables: - $INTNAME1 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $INTNAME1 - TableScan - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n6 - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) - outputColumnNames: key, cnt1, cnt2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt1, 'hll'), compute_stats(cnt2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n6 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt1, cnt2 - Column Types: string, int, int - Table: default.dest_j1_n6 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-10 - Map Reduce Local Work - Alias -> Map Local Tables: - $INTNAME - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $INTNAME - TableScan - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n6 - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) - outputColumnNames: key, cnt1, cnt2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt1, 'hll'), compute_stats(cnt2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col3) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n6 - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) - outputColumnNames: key, cnt1, cnt2 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt1, 'hll'), compute_stats(cnt2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n6 -SELECT subq1.key, subq1.cnt, subq2.cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n6 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n6 -SELECT subq1.key, subq1.cnt, subq2.cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n6 -POSTHOOK: Lineage: dest_j1_n6.cnt1 EXPRESSION [(src1)x.null, ] -POSTHOOK: Lineage: dest_j1_n6.cnt2 EXPRESSION [(src)y.null, ] -POSTHOOK: Lineage: dest_j1_n6.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n6 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n6 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n6 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n6 -#### A masked pattern was here #### -128 1 3 -146 1 2 -150 1 1 -213 1 2 -224 1 2 -238 1 2 -255 1 2 -273 1 3 -278 1 2 -311 1 3 -369 1 3 -401 1 5 -406 1 4 -66 1 1 -98 1 2 diff --git ql/src/test/results/clientpositive/join30.q.out ql/src/test/results/clientpositive/join30.q.out deleted file mode 100644 index 6df2aab26d..0000000000 --- ql/src/test/results/clientpositive/join30.q.out +++ /dev/null @@ -1,208 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n0(key INT, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n0 -POSTHOOK: query: CREATE TABLE dest_j1_n0(key INT, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n0 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n0 -SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n0 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n0 -SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n0 -STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-2 depends on stages: Stage-7 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-7 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 39 Data size: 3354 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Local Work: - Map Reduce Local Work - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n0 - Select Operator - expressions: _col0 (type: int), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n0 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: int, int - Table: default.dest_j1_n0 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n0 -SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n0 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n0 -SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n0 -POSTHOOK: Lineage: dest_j1_n0.cnt EXPRESSION [(src1)x.null, (src)y.null, ] -POSTHOOK: Lineage: dest_j1_n0.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n0 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n0 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n0 -#### A masked pattern was here #### -128 3 -146 2 -150 1 -213 2 -224 2 -238 2 -255 2 -273 3 -278 2 -311 3 -369 3 -401 5 -406 4 -66 1 -98 2 diff --git ql/src/test/results/clientpositive/join31.q.out ql/src/test/results/clientpositive/join31.q.out deleted file mode 100644 index c0455a2a2e..0000000000 --- ql/src/test/results/clientpositive/join31.q.out +++ /dev/null @@ -1,253 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n22(key STRING, cnt INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n22 -POSTHOOK: query: CREATE TABLE dest_j1_n22(key STRING, cnt INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n22 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n22 -SELECT subq1.key, count(1) as cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -group by subq1.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n22 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n22 -SELECT subq1.key, count(1) as cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -group by subq1.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n22 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-8 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-8 - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0, Stage-5 - Stage-5 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-8 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:y - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:y - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Map Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Local Work: - Map Reduce Local Work - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n22 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n22 - - Stage: Stage-4 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, cnt - Column Types: string, int - Table: default.dest_j1_n22 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n22 -SELECT subq1.key, count(1) as cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -group by subq1.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n22 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n22 -SELECT subq1.key, count(1) as cnt -FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN - (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) -group by subq1.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n22 -POSTHOOK: Lineage: dest_j1_n22.cnt EXPRESSION [(src1)x.null, ] -POSTHOOK: Lineage: dest_j1_n22.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n22 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n22 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n22 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n22 -#### A masked pattern was here #### -128 1 -146 1 -150 1 -213 1 -224 1 -238 1 -255 1 -273 1 -278 1 -311 1 -369 1 -401 1 -406 1 -66 1 -98 1 diff --git ql/src/test/results/clientpositive/join32.q.out ql/src/test/results/clientpositive/join32.q.out deleted file mode 100644 index a8b9ae9e74..0000000000 --- ql/src/test/results/clientpositive/join32.q.out +++ /dev/null @@ -1,629 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n12(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n12 -POSTHOOK: query: CREATE TABLE dest_j1_n12(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n12 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n12 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n12 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n12 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n12 -OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` -FROM (SELECT `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `value` IS NOT NULL) AS `t0` -INNER JOIN ((SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` IS NOT NULL) AS `t2` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src1` -WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`value` = `t4`.`value` -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-7 depends on stages: Stage-9 - Stage-0 depends on stages: Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:z - Fetch Operator - limit: -1 - Partition Description: - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - $hdt$_2:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:z - TableScan - alias: z - filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D) and value is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col3 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - $hdt$_2:x - TableScan - alias: x - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col3 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col4 - Position of Big Table: 0 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n12 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n12 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n12 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Local Work: - Map Reduce Local Work - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: src1 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src1 - name: default.src1 -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:y] - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n12 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n12 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n12 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, val2 - Column Types: string, string, string - Table: default.dest_j1_n12 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n12 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n12 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n12 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n12 -POSTHOOK: Lineage: dest_j1_n12.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n12.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n12.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n12 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n12 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n12 -#### A masked pattern was here #### -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 diff --git ql/src/test/results/clientpositive/join33.q.out ql/src/test/results/clientpositive/join33.q.out deleted file mode 100644 index b53a32ac1c..0000000000 --- ql/src/test/results/clientpositive/join33.q.out +++ /dev/null @@ -1,629 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n7(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n7 -POSTHOOK: query: CREATE TABLE dest_j1_n7(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n7 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n7 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n7 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n7 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n7 -OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` -FROM (SELECT `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `value` IS NOT NULL) AS `t0` -INNER JOIN ((SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` IS NOT NULL) AS `t2` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src1` -WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`value` = `t4`.`value` -STAGE DEPENDENCIES: - Stage-9 is a root stage - Stage-7 depends on stages: Stage-9 - Stage-0 depends on stages: Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-7 - -STAGE PLANS: - Stage: Stage-9 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:z - Fetch Operator - limit: -1 - Partition Description: - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - $hdt$_2:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:z - TableScan - alias: z - filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D) and value is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col3 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - $hdt$_2:x - TableScan - alias: x - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col3 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col4 - Position of Big Table: 0 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n7 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n7 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Local Work: - Map Reduce Local Work - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: src1 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src1 - name: default.src1 -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:y] - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n7 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n7 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, val2 - Column Types: string, string, string - Table: default.dest_j1_n7 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n7 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Output: default@dest_j1_n7 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n7 -SELECT x.key, z.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Output: default@dest_j1_n7 -POSTHOOK: Lineage: dest_j1_n7.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n7.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n7.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n7 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n7 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n7 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n7 -#### A masked pattern was here #### -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -213 val_213 val_213 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 -98 val_98 val_98 diff --git ql/src/test/results/clientpositive/join34.q.out ql/src/test/results/clientpositive/join34.q.out deleted file mode 100644 index fdeb0227ba..0000000000 --- ql/src/test/results/clientpositive/join34.q.out +++ /dev/null @@ -1,551 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n1 -POSTHOOK: query: CREATE TABLE dest_j1_n1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n1 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n1 -SELECT x.key, x.value, subq1.value -FROM -( SELECT x.key as key, x.value as value from src x where x.key < 20 - UNION ALL - SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 -) subq1 -JOIN src1 x ON (x.key = subq1.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n1 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n1 -SELECT x.key, x.value, subq1.value -FROM -( SELECT x.key as key, x.value as value from src x where x.key < 20 - UNION ALL - SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 -) subq1 -JOIN src1 x ON (x.key = subq1.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n1 -OPTIMIZED SQL: SELECT `t5`.`key`, `t5`.`value`, `t3`.`value` AS `value1` -FROM (SELECT `key`, `value` -FROM `default`.`src` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` < 20 -UNION ALL -SELECT `key`, `value` -FROM `default`.`src` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` > 100) AS `t3` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src1` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` IS NOT NULL) AS `t5` ON `t3`.`key` = `t5`.`key` -STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-6 depends on stages: Stage-7 - Stage-0 depends on stages: Stage-6 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-6 - -STAGE PLANS: - Stage: Stage-7 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:x - TableScan - alias: x - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) - Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 296 Data size: 52688 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col3 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - TableScan - alias: x1 - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) - Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 296 Data size: 52688 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col3 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Local Work: - Map Reduce Local Work - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: src1 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src1 - name: default.src1 - Truncated Path -> Alias: - /src [$hdt$_0:$hdt$_0-subquery1:x, $hdt$_0:$hdt$_0-subquery2:x1] - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:string -#### A masked pattern was here #### - name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n1 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, val2 - Column Types: string, string, string - Table: default.dest_j1_n1 - Is Table Level Stats: true - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n1 -SELECT x.key, x.value, subq1.value -FROM -( SELECT x.key as key, x.value as value from src x where x.key < 20 - UNION ALL - SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 -) subq1 -JOIN src1 x ON (x.key = subq1.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n1 -SELECT x.key, x.value, subq1.value -FROM -( SELECT x.key as key, x.value as value from src x where x.key < 20 - UNION ALL - SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 -) subq1 -JOIN src1 x ON (x.key = subq1.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n1 -POSTHOOK: Lineage: dest_j1_n1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n1.val2 EXPRESSION [(src)x.FieldSchema(name:value, type:string, comment:default), (src)x1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n1.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n1 -#### A masked pattern was here #### -128 val_128 -128 val_128 -128 val_128 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -224 val_224 -224 val_224 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -369 val_369 -369 val_369 -369 val_369 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 diff --git ql/src/test/results/clientpositive/join35.q.out ql/src/test/results/clientpositive/join35.q.out deleted file mode 100644 index f4d2c0cfb9..0000000000 --- ql/src/test/results/clientpositive/join35.q.out +++ /dev/null @@ -1,732 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n24(key STRING, value STRING, val2 INT) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n24 -POSTHOOK: query: CREATE TABLE dest_j1_n24(key STRING, value STRING, val2 INT) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n24 -PREHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n24 -SELECT x.key, x.value, subq1.cnt -FROM -( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key - UNION ALL - SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key -) subq1 -JOIN src1 x ON (x.key = subq1.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n24 -POSTHOOK: query: EXPLAIN EXTENDED -INSERT OVERWRITE TABLE dest_j1_n24 -SELECT x.key, x.value, subq1.cnt -FROM -( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key - UNION ALL - SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key -) subq1 -JOIN src1 x ON (x.key = subq1.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n24 -OPTIMIZED SQL: SELECT `t5`.`key`, `t5`.`value`, `t3`.`$f1` AS `cnt` -FROM (SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`src` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` < 20 -GROUP BY `key` -UNION ALL -SELECT `key`, COUNT(*) AS `$f1` -FROM `default`.`src` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` > 100 -GROUP BY `key`) AS `t3` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src1` -WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` IS NOT NULL) AS `t5` ON `t3`.`key` = `t5`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-8 depends on stages: Stage-1, Stage-5 - Stage-7 depends on stages: Stage-8 - Stage-0 depends on stages: Stage-7 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-7 - Stage-5 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 148 Data size: 12876 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [$hdt$_0:$hdt$_0-subquery1:x] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-8 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:x - TableScan - alias: x - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) - Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Position of Big Table: 0 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Union - Statistics: Num rows: 148 Data size: 14060 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 46 Data size: 8418 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col3 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:int -#### A masked pattern was here #### - name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n24 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, value, val2 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - TableScan - GatherStats: false - Union - Statistics: Num rows: 148 Data size: 14060 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Position of Big Table: 0 - Statistics: Num rows: 46 Data size: 8418 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col3 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:int -#### A masked pattern was here #### - name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n24 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, value, val2 - Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Local Work: - Map Reduce Local Work - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: src1 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src1 - name: default.src1 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value,val2 - columns.comments - columns.types string:string:int -#### A masked pattern was here #### - name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n24 - - Stage: Stage-3 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value, val2 - Column Types: string, string, int - Table: default.dest_j1_n24 - Is Table Level Stats: true - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10003 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2 - columns.types struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: x1 - filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) - Statistics: Num rows: 148 Data size: 12876 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [$hdt$_0:$hdt$_0-subquery2:x1] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types string,bigint - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n24 -SELECT x.key, x.value, subq1.cnt -FROM -( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key - UNION ALL - SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key -) subq1 -JOIN src1 x ON (x.key = subq1.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n24 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n24 -SELECT x.key, x.value, subq1.cnt -FROM -( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key - UNION ALL - SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key -) subq1 -JOIN src1 x ON (x.key = subq1.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n24 -POSTHOOK: Lineage: dest_j1_n24.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n24.val2 EXPRESSION [(src)x.null, (src)x1.null, ] -POSTHOOK: Lineage: dest_j1_n24.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n24 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n24 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n24 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n24 -#### A masked pattern was here #### -128 3 -146 val_146 2 -150 val_150 1 -213 val_213 2 -224 2 -238 val_238 2 -255 val_255 2 -273 val_273 3 -278 val_278 2 -311 val_311 3 -369 3 -401 val_401 5 -406 val_406 4 diff --git ql/src/test/results/clientpositive/join37.q.out ql/src/test/results/clientpositive/join37.q.out deleted file mode 100644 index 9d0709623f..0000000000 --- ql/src/test/results/clientpositive/join37.q.out +++ /dev/null @@ -1,213 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_j1_n9(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_j1_n9 -POSTHOOK: query: CREATE TABLE dest_j1_n9(key INT, value STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_j1_n9 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n9 -SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n9 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_j1_n9 -SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n9 -STAGE DEPENDENCIES: - Stage-6 is a root stage - Stage-5 depends on stages: Stage-6 - Stage-0 depends on stages: Stage-5 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-6 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n9 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, val2 - Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n9 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value, val2 - Column Types: int, string, string - Table: default.dest_j1_n9 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n9 -SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_j1_n9 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n9 -SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value -FROM src1 x JOIN src y ON (x.key = y.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_j1_n9 -POSTHOOK: Lineage: dest_j1_n9.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n9.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest_j1_n9.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: select * from dest_j1_n9 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest_j1_n9 -#### A masked pattern was here #### -POSTHOOK: query: select * from dest_j1_n9 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest_j1_n9 -#### A masked pattern was here #### -128 val_128 -128 val_128 -128 val_128 -146 val_146 val_146 -146 val_146 val_146 -150 val_150 val_150 -213 val_213 val_213 -213 val_213 val_213 -224 val_224 -224 val_224 -238 val_238 val_238 -238 val_238 val_238 -255 val_255 val_255 -255 val_255 val_255 -273 val_273 val_273 -273 val_273 val_273 -273 val_273 val_273 -278 val_278 val_278 -278 val_278 val_278 -311 val_311 val_311 -311 val_311 val_311 -311 val_311 val_311 -369 val_369 -369 val_369 -369 val_369 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -401 val_401 val_401 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -406 val_406 val_406 -66 val_66 val_66 -98 val_98 val_98 -98 val_98 val_98 diff --git ql/src/test/results/clientpositive/join38.q.out ql/src/test/results/clientpositive/join38.q.out deleted file mode 100644 index 38341bb8d8..0000000000 --- ql/src/test/results/clientpositive/join38.q.out +++ /dev/null @@ -1,163 +0,0 @@ -PREHOOK: query: create table tmp_n1(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tmp_n1 -POSTHOOK: query: create table tmp_n1(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tmp_n1 -PREHOOK: query: insert overwrite table tmp_n1 select key, cast(key + 1 as int), key +2, key+3, key+4, cast(key+5 as int), key+6, key+7, key+8, key+9, key+10, cast(key+11 as int) from src where key = 100 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tmp_n1 -POSTHOOK: query: insert overwrite table tmp_n1 select key, cast(key + 1 as int), key +2, key+3, key+4, cast(key+5 as int), key+6, key+7, key+8, key+9, key+10, cast(key+11 as int) from src where key = 100 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tmp_n1 -POSTHOOK: Lineage: tmp_n1.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: tmp_n1.col1 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col10 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col11 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col2 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col3 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col4 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col5 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col6 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col7 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col8 EXPRESSION [] -POSTHOOK: Lineage: tmp_n1.col9 EXPRESSION [] -PREHOOK: query: select * from tmp_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -POSTHOOK: query: select * from tmp_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -100 101 102.0 103.0 104.0 105 106.0 107.0 108.0 109.0 110.0 111 -100 101 102.0 103.0 104.0 105 106.0 107.0 108.0 109.0 110.0 111 -PREHOOK: query: explain -FROM src a JOIN tmp_n1 b ON (a.key = b.col11) -SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count -where b.col11 = 111 -group by a.value, b.col5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -POSTHOOK: query: explain -FROM src a JOIN tmp_n1 b ON (a.key = b.col11) -SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count -where b.col11 = 111 -group by a.value, b.col5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-5 is a root stage - Stage-2 depends on stages: Stage-5 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-5 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:b - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:b - TableScan - alias: b - filterExpr: (UDFToDouble(col11) = 111.0D) (type: boolean) - Statistics: Num rows: 2 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(col11) = 111.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: col5 (type: string), col11 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col1 (type: string) - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: (UDFToDouble(key) = 111.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) = 111.0D) (type: boolean) - Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: _col1 (type: string), _col2 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Local Work: - Map Reduce Local Work - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM src a JOIN tmp_n1 b ON (a.key = b.col11) -SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count -where b.col11 = 111 -group by a.value, b.col5 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -POSTHOOK: query: FROM src a JOIN tmp_n1 b ON (a.key = b.col11) -SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count -where b.col11 = 111 -group by a.value, b.col5 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@tmp_n1 -#### A masked pattern was here #### -val_111 105 2 diff --git ql/src/test/results/clientpositive/join4.q.out ql/src/test/results/clientpositive/join4.q.out deleted file mode 100644 index 3515bf3634..0000000000 --- ql/src/test/results/clientpositive/join4.q.out +++ /dev/null @@ -1,223 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n72(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n72 -POSTHOOK: query: CREATE TABLE dest1_n72(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n72 -PREHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n72 -POSTHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n72 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n72 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) - outputColumnNames: c1, c2, c3, c4 - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n72 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3, c4 - Column Types: int, string, int, string - Table: default.dest1_n72 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n72 -POSTHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n72 -POSTHOOK: Lineage: dest1_n72.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n72.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n72.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n72.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n72.* FROM dest1_n72 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n72 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n72.* FROM dest1_n72 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n72 -#### A masked pattern was here #### -11 val_11 NULL NULL -12 val_12 NULL NULL -12 val_12 NULL NULL -15 val_15 NULL NULL -15 val_15 NULL NULL -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 diff --git ql/src/test/results/clientpositive/join41.q.out ql/src/test/results/clientpositive/join41.q.out deleted file mode 100644 index 744401680b..0000000000 --- ql/src/test/results/clientpositive/join41.q.out +++ /dev/null @@ -1,180 +0,0 @@ -PREHOOK: query: create table s1 as select * from src where key = 0 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@s1 -POSTHOOK: query: create table s1 as select * from src where key = 0 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@s1 -POSTHOOK: Lineage: s1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: s1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: EXPLAIN -SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) > 10.0D) (type: boolean) - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) > 10.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -0 val_0 NULL NULL -0 val_0 NULL NULL -0 val_0 NULL NULL -PREHOOK: query: EXPLAIN -SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) > 10.0D) (type: boolean) - Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) > 10.0D) (type: boolean) - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -PREHOOK: type: QUERY -PREHOOK: Input: default@s1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@s1 -#### A masked pattern was here #### -0 val_0 NULL NULL -0 val_0 NULL NULL -0 val_0 NULL NULL diff --git ql/src/test/results/clientpositive/join42.q.out ql/src/test/results/clientpositive/join42.q.out deleted file mode 100644 index a32e0b4d98..0000000000 --- ql/src/test/results/clientpositive/join42.q.out +++ /dev/null @@ -1,415 +0,0 @@ -PREHOOK: query: create table L as select 4436 id -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@L -POSTHOOK: query: create table L as select 4436 id -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@L -POSTHOOK: Lineage: l.id SIMPLE [] -PREHOOK: query: create table LA_n11 as select 4436 loan_id, 4748 aid, 4415 pi_id -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@LA_n11 -POSTHOOK: query: create table LA_n11 as select 4436 loan_id, 4748 aid, 4415 pi_id -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@LA_n11 -POSTHOOK: Lineage: la_n11.aid SIMPLE [] -POSTHOOK: Lineage: la_n11.loan_id SIMPLE [] -POSTHOOK: Lineage: la_n11.pi_id SIMPLE [] -PREHOOK: query: create table FR as select 4436 loan_id -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@FR -POSTHOOK: query: create table FR as select 4436 loan_id -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@FR -POSTHOOK: Lineage: fr.loan_id SIMPLE [] -PREHOOK: query: create table A_n11 as select 4748 id -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@A_n11 -POSTHOOK: query: create table A_n11 as select 4748 id -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@A_n11 -POSTHOOK: Lineage: a_n11.id SIMPLE [] -PREHOOK: query: create table PI as select 4415 id -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@PI -POSTHOOK: query: create table PI as select 4415 id -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@PI -POSTHOOK: Lineage: pi.id SIMPLE [] -PREHOOK: query: create table acct as select 4748 aid, 10 acc_n, 122 brn -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: database:default -PREHOOK: Output: default@acct -POSTHOOK: query: create table acct as select 4748 aid, 10 acc_n, 122 brn -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: database:default -POSTHOOK: Output: default@acct -POSTHOOK: Lineage: acct.acc_n SIMPLE [] -POSTHOOK: Lineage: acct.aid SIMPLE [] -POSTHOOK: Lineage: acct.brn SIMPLE [] -PREHOOK: query: insert into table acct values(4748, null, null) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@acct -POSTHOOK: query: insert into table acct values(4748, null, null) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@acct -POSTHOOK: Lineage: acct.acc_n EXPRESSION [] -POSTHOOK: Lineage: acct.aid SCRIPT [] -POSTHOOK: Lineage: acct.brn EXPRESSION [] -PREHOOK: query: insert into table acct values(4748, null, null) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@acct -POSTHOOK: query: insert into table acct values(4748, null, null) -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@acct -POSTHOOK: Lineage: acct.acc_n EXPRESSION [] -POSTHOOK: Lineage: acct.aid SCRIPT [] -POSTHOOK: Lineage: acct.brn EXPRESSION [] -Warning: Shuffle Join JOIN[30][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_4, $hdt$_5]] in Stage 'Stage-8:MAPRED' is a cross product -PREHOOK: query: explain select - acct.ACC_N, - acct.brn -FROM L -JOIN LA_n11 ON L.id = LA_n11.loan_id -JOIN FR ON L.id = FR.loan_id -JOIN A_n11 ON LA_n11.aid = A_n11.id -JOIN PI ON PI.id = LA_n11.pi_id -JOIN acct ON A_n11.id = acct.aid -WHERE - L.id = 4436 - and acct.brn is not null -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n11 -PREHOOK: Input: default@acct -PREHOOK: Input: default@fr -PREHOOK: Input: default@l -PREHOOK: Input: default@la_n11 -PREHOOK: Input: default@pi -#### A masked pattern was here #### -POSTHOOK: query: explain select - acct.ACC_N, - acct.brn -FROM L -JOIN LA_n11 ON L.id = LA_n11.loan_id -JOIN FR ON L.id = FR.loan_id -JOIN A_n11 ON LA_n11.aid = A_n11.id -JOIN PI ON PI.id = LA_n11.pi_id -JOIN acct ON A_n11.id = acct.aid -WHERE - L.id = 4436 - and acct.brn is not null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n11 -POSTHOOK: Input: default@acct -POSTHOOK: Input: default@fr -POSTHOOK: Input: default@l -POSTHOOK: Input: default@la_n11 -POSTHOOK: Input: default@pi -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-3, Stage-8 - Stage-8 is a root stage - Stage-0 depends on stages: Stage-4 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: la_n11 - filterExpr: ((loan_id = 4436) and aid is not null and pi_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((loan_id = 4436) and aid is not null and pi_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: aid (type: int), pi_id (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: a_n11 - filterExpr: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: id (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int) - TableScan - alias: pi - filterExpr: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: id (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col2 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int) - TableScan - alias: fr - filterExpr: (loan_id = 4436) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (loan_id = 4436) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col2 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: int) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int), _col2 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col6, _col7 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col6 (type: int), _col7 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-8 - Map Reduce - Map Operator Tree: - TableScan - alias: acct - filterExpr: (brn is not null and aid is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (brn is not null and aid is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: aid (type: int), acc_n (type: int), brn (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - TableScan - alias: l - filterExpr: (id = 4436) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (id = 4436) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[30][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_4, $hdt$_5]] in Stage 'Stage-8:MAPRED' is a cross product -PREHOOK: query: select - acct.ACC_N, - acct.brn -FROM L -JOIN LA_n11 ON L.id = LA_n11.loan_id -JOIN FR ON L.id = FR.loan_id -JOIN A_n11 ON LA_n11.aid = A_n11.id -JOIN PI ON PI.id = LA_n11.pi_id -JOIN acct ON A_n11.id = acct.aid -WHERE - L.id = 4436 - and acct.brn is not null -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n11 -PREHOOK: Input: default@acct -PREHOOK: Input: default@fr -PREHOOK: Input: default@l -PREHOOK: Input: default@la_n11 -PREHOOK: Input: default@pi -#### A masked pattern was here #### -POSTHOOK: query: select - acct.ACC_N, - acct.brn -FROM L -JOIN LA_n11 ON L.id = LA_n11.loan_id -JOIN FR ON L.id = FR.loan_id -JOIN A_n11 ON LA_n11.aid = A_n11.id -JOIN PI ON PI.id = LA_n11.pi_id -JOIN acct ON A_n11.id = acct.aid -WHERE - L.id = 4436 - and acct.brn is not null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n11 -POSTHOOK: Input: default@acct -POSTHOOK: Input: default@fr -POSTHOOK: Input: default@l -POSTHOOK: Input: default@la_n11 -POSTHOOK: Input: default@pi -#### A masked pattern was here #### -10 122 diff --git ql/src/test/results/clientpositive/join44.q.out ql/src/test/results/clientpositive/join44.q.out deleted file mode 100644 index c2110043c4..0000000000 --- ql/src/test/results/clientpositive/join44.q.out +++ /dev/null @@ -1,92 +0,0 @@ -PREHOOK: query: CREATE TABLE mytable_n1(val1 INT, val2 INT, val3 INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@mytable_n1 -POSTHOOK: query: CREATE TABLE mytable_n1(val1 INT, val2 INT, val3 INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@mytable_n1 -PREHOOK: query: EXPLAIN -SELECT * -FROM mytable_n1 src1, mytable_n1 src2 -WHERE src1.val1=src2.val1 - AND src1.val2 between 2450816 and 2451500 - AND src2.val2 between 2450816 and 2451500 -PREHOOK: type: QUERY -PREHOOK: Input: default@mytable_n1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM mytable_n1 src1, mytable_n1 src2 -WHERE src1.val1=src2.val1 - AND src1.val2 between 2450816 and 2451500 - AND src2.val2 between 2450816 and 2451500 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@mytable_n1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: val1 (type: int) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - value expressions: val2 (type: int), val3 (type: int) - TableScan - alias: src2 - filterExpr: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: val1 (type: int) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - value expressions: val2 (type: int), val3 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 val1 (type: int) - 1 val1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (_col0 = _col6) (type: boolean) - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join45.q.out ql/src/test/results/clientpositive/join45.q.out deleted file mode 100644 index 542ce5f73f..0000000000 --- ql/src/test/results/clientpositive/join45.q.out +++ /dev/null @@ -1,1979 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -128 128 val_128 -128 128 val_128 -128 128 val_128 -146 val_146 146 val_146 -146 val_146 146 val_146 -150 val_150 150 val_150 -213 val_213 213 val_213 -213 val_213 213 val_213 -224 224 val_224 -224 224 val_224 -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col0 = _col3) or _col2 or _col5) (type: boolean) - Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -146 val_146 146 val_146 -146 val_146 146 val_146 -213 val_213 213 val_213 -213 val_213 213 val_213 -128 128 val_128 -128 128 val_128 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4612500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col2 + _col5) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -66 val_66 35 val_35 -66 val_66 34 val_34 -66 val_66 35 val_35 -66 val_66 35 val_35 -98 val_98 2 val_2 -98 val_98 4 val_4 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4612500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((_col2 + _col5) >= 100.0D) or ((_col2 + _col5) <= 102.0D)) (type: boolean) - Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -406 val_406 97 val_97 -406 val_406 200 val_200 -406 val_406 400 val_400 -406 val_406 403 val_403 -406 val_406 169 val_169 -406 val_406 90 val_90 -406 val_406 126 val_126 -406 val_406 222 val_222 -406 val_406 477 val_477 -406 val_406 414 val_414 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4612500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (struct(_col2,_col5)) IN (const struct(100.0D,100.0D), const struct(101.0D,101.0D), const struct(102.0D,102.0D)) (type: boolean) - Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 13475 Data size: 7243075 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col7 + _col2) >= 100.0D) (type: boolean) - Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -98 val_98 406 val_406 98 val_98 -98 val_98 146 val_146 98 val_98 -98 val_98 213 val_213 98 val_98 -98 val_98 128 98 val_98 -98 val_98 66 val_66 98 val_98 -98 val_98 369 98 val_98 -98 val_98 224 98 val_98 -98 val_98 273 val_273 98 val_98 -98 val_98 150 val_150 98 val_98 -98 val_98 401 val_401 98 val_98 -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 13475 Data size: 7243075 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col7 + _col2) <= 102.0D) (type: boolean) - Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -9 val_9 NULL NULL 66 val_66 -8 val_8 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -4 val_4 NULL NULL 66 val_66 -4 val_4 NULL NULL 98 val_98 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4612500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col5 + _col2) >= 100.0D) (type: boolean) - Statistics: Num rows: 4166 Data size: 1537254 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL NULL NULL val_484 -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL val_27 -NULL NULL NULL NULL val_165 -NULL NULL NULL NULL val_409 -NULL NULL NULL NULL -NULL NULL NULL NULL val_193 -NULL NULL NULL NULL val_265 -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1600 Data size: 865936 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col7 + _col2) <= 102.0D) (type: boolean) - Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 533 Data size: 280178 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 12500 Data size: 4612500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col5 + _col2) >= 100.0D) (type: boolean) - Statistics: Num rows: 4166 Data size: 1537254 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL NULL NULL val_484 -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL val_27 -NULL NULL NULL NULL val_165 -NULL NULL NULL NULL val_409 -NULL NULL NULL NULL -NULL NULL NULL NULL val_193 -NULL NULL NULL NULL val_265 -Warning: Shuffle Join JOIN[11][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 14100 Data size: 7481236 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col7 + _col2) <= 102.0D) (type: boolean) - Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4700 Data size: 2420178 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[11][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -9 val_9 NULL NULL 66 val_66 -8 val_8 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -4 val_4 NULL NULL 66 val_66 -4 val_4 NULL NULL 98 val_98 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -Warning: Shuffle Join JOIN[23][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -Warning: Shuffle Join JOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@cbo_t1 -PREHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@cbo_t1 -POSTHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2, Stage-6 - Stage-6 is a root stage - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string), (c_int = 42) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean) - TableScan - alias: t3 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string) - TableScan - alias: t4 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18 - Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string) - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col11 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30 - Statistics: Num rows: 720000 Data size: 1320479175 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((_col7 = _col25) or ((_col8 = _col26) and (_col27 = _col21))) and ((_col13 = _col25) or ((_col28 = _col16) and (_col14 = _col26))) and ((_col0 = _col25) or ((_col29 = _col17) and _col6))) (type: boolean) - Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col30 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 - Statistics: Num rows: 303750 Data size: 555862500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - alias: t1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - TableScan - alias: t2 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[23][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -Warning: Shuffle Join JOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@cbo_t1 -PREHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@cbo_t1 -POSTHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 diff --git ql/src/test/results/clientpositive/join47.q.out ql/src/test/results/clientpositive/join47.q.out deleted file mode 100644 index 989489d0b6..0000000000 --- ql/src/test/results/clientpositive/join47.q.out +++ /dev/null @@ -1,1957 +0,0 @@ -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - AND src1.value between 100 and 102 - AND src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -128 128 val_128 -128 128 val_128 -128 128 val_128 -146 val_146 146 val_146 -146 val_146 146 val_146 -150 val_150 150 val_150 -213 val_213 213 val_213 -213 val_213 213 val_213 -224 224 val_224 -224 224 val_224 -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.value between 100 and 102 - AND src.value between 100 and 102 - AND true) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {((_col0 = _col3) or _col2 or _col5)} - Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key=src.key - OR src1.value between 100 and 102 - OR src.value between 100 and 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -406 val_406 406 val_406 -146 val_146 146 val_146 -146 val_146 146 val_146 -213 val_213 213 val_213 -213 val_213 213 val_213 -128 128 val_128 -128 128 val_128 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {(_col2 + _col5) BETWEEN 100.0D AND 102.0D} - Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - AND src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -66 val_66 35 val_35 -66 val_66 34 val_34 -66 val_66 35 val_35 -66 val_66 35 val_35 -98 val_98 2 val_2 -98 val_98 4 val_4 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {(((_col2 + _col5) >= 100.0D) or ((_col2 + _col5) <= 102.0D))} - Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON (src1.key+src.key >= 100 - OR src1.key+src.key <= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -406 val_406 97 val_97 -406 val_406 200 val_200 -406 val_406 400 val_400 -406 val_406 403 val_403 -406 val_406 169 val_169 -406 val_406 90 val_90 -406 val_406 126 val_126 -406 val_406 222 val_222 -406 val_406 477 val_477 -406 val_406 414 val_414 -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {(struct(_col2,_col5)) IN (const struct(100.0D,100.0D), const struct(101.0D,101.0D), const struct(102.0D,102.0D))} - Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src1 JOIN src -ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - residual filter predicates: {((_col7 + _col2) >= 100.0D)} - Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -LEFT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -98 val_98 406 val_406 98 val_98 -98 val_98 146 val_146 98 val_98 -98 val_98 213 val_213 98 val_98 -98 val_98 128 98 val_98 -98 val_98 66 val_66 98 val_98 -98 val_98 369 98 val_98 -98 val_98 224 98 val_98 -98 val_98 273 val_273 98 val_98 -98 val_98 150 val_150 98 val_98 -98 val_98 401 val_401 98 val_98 -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - residual filter predicates: {((_col7 + _col2) <= 102.0D)} - Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -LEFT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -9 val_9 NULL NULL 66 val_66 -8 val_8 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -4 val_4 NULL NULL 66 val_66 -4 val_4 NULL NULL 98 val_98 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {((_col5 + _col2) >= 100.0D)} - Statistics: Num rows: 4166 Data size: 1537254 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[7][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -RIGHT OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL NULL NULL val_484 -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL val_27 -NULL NULL NULL NULL val_165 -NULL NULL NULL NULL val_409 -NULL NULL NULL NULL -NULL NULL NULL NULL val_193 -NULL NULL NULL NULL val_265 -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - residual filter predicates: {((_col7 + _col2) <= 102.0D)} - Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 533 Data size: 280178 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[12][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -RIGHT OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - residual filter predicates: {((_col5 + _col2) >= 100.0D)} - Statistics: Num rows: 4166 Data size: 1537254 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -JOIN src1 a ON (a.key+src.key >= 100) -FULL OUTER JOIN src1 b ON (b.key = src.key) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -NULL NULL NULL NULL val_484 -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL -NULL NULL NULL NULL val_27 -NULL NULL NULL NULL val_165 -NULL NULL NULL NULL val_409 -NULL NULL NULL NULL -NULL NULL NULL NULL val_193 -NULL NULL NULL NULL val_265 -Warning: Shuffle Join JOIN[11][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: double) - TableScan - alias: a - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - residual filter predicates: {((_col7 + _col2) <= 102.0D)} - Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 4700 Data size: 2420178 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[11][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM src -FULL OUTER JOIN src1 a ON (a.key = src.key) -JOIN src1 b ON (b.key+src.key<= 102) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -9 val_9 NULL NULL 66 val_66 -8 val_8 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -5 val_5 NULL NULL 66 val_66 -4 val_4 NULL NULL 66 val_66 -4 val_4 NULL NULL 98 val_98 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -35 val_35 NULL NULL 66 val_66 -Warning: Shuffle Join JOIN[23][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -Warning: Shuffle Join JOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN -SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@cbo_t1 -PREHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@cbo_t1 -POSTHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2, Stage-6 - Stage-6 is a root stage - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string), (c_int = 42) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean) - TableScan - alias: t3 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string) - TableScan - alias: t4 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18 - Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string) - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col11 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30 - residual filter predicates: {((_col0 = _col25) or ((_col29 = _col17) and _col6))} {((_col13 = _col25) or ((_col28 = _col16) and (_col14 = _col26)))} {((_col7 = _col25) or ((_col8 = _col26) and (_col27 = _col21)))} - Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col30 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 - Statistics: Num rows: 303750 Data size: 555862500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - alias: t1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - TableScan - alias: t2 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[23][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Stage-3:MAPRED' is a cross product -Warning: Shuffle Join JOIN[20][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -Warning: Shuffle Join JOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@cbo_t1 -PREHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -POSTHOOK: query: SELECT * -FROM cbo_t1 t1 -RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) -JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) -JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) -JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) -LIMIT 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@cbo_t1 -POSTHOOK: Input: default@cbo_t1@dt=2014 -#### A masked pattern was here #### -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 -1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 false 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 diff --git ql/src/test/results/clientpositive/join5.q.out ql/src/test/results/clientpositive/join5.q.out deleted file mode 100644 index c91d5a60eb..0000000000 --- ql/src/test/results/clientpositive/join5.q.out +++ /dev/null @@ -1,221 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n126(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n126 -POSTHOOK: query: CREATE TABLE dest1_n126(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n126 -PREHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - RIGHT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n126 -POSTHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - RIGHT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n126 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n126 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) - outputColumnNames: c1, c2, c3, c4 - Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n126 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3, c4 - Column Types: int, string, int, string - Table: default.dest1_n126 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - RIGHT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n126 -POSTHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - RIGHT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n126 -POSTHOOK: Lineage: dest1_n126.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n126.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n126.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n126.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n126.* FROM dest1_n126 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n126 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n126.* FROM dest1_n126 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n126 -#### A masked pattern was here #### -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -NULL NULL 20 val_20 -NULL NULL 24 val_24 -NULL NULL 24 val_24 diff --git ql/src/test/results/clientpositive/join6.q.out ql/src/test/results/clientpositive/join6.q.out deleted file mode 100644 index e0fd2aefb0..0000000000 --- ql/src/test/results/clientpositive/join6.q.out +++ /dev/null @@ -1,226 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n156(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n156 -POSTHOOK: query: CREATE TABLE dest1_n156(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n156 -PREHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n156 -POSTHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n156 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n156 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) - outputColumnNames: c1, c2, c3, c4 - Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n156 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3, c4 - Column Types: int, string, int, string - Table: default.dest1_n156 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n156 -POSTHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n156 -POSTHOOK: Lineage: dest1_n156.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n156.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n156.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n156.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n156.* FROM dest1_n156 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n156 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n156.* FROM dest1_n156 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n156 -#### A masked pattern was here #### -11 val_11 NULL NULL -12 val_12 NULL NULL -12 val_12 NULL NULL -15 val_15 NULL NULL -15 val_15 NULL NULL -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -NULL NULL 20 val_20 -NULL NULL 24 val_24 -NULL NULL 24 val_24 diff --git ql/src/test/results/clientpositive/join7.q.out ql/src/test/results/clientpositive/join7.q.out deleted file mode 100644 index b970bd7596..0000000000 --- ql/src/test/results/clientpositive/join7.q.out +++ /dev/null @@ -1,294 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n17(c1 INT, c2 STRING, c3 INT, c4 STRING, c5 INT, c6 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n17 -POSTHOOK: query: CREATE TABLE dest1_n17(c1 INT, c2 STRING, c3 INT, c4 STRING, c5 INT, c6 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n17 -PREHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - LEFT OUTER JOIN - ( - FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 - ) c - ON (a.c1 = c.c5) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 -) c -INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n17 -POSTHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - LEFT OUTER JOIN - ( - FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 - ) c - ON (a.c1 = c.c5) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 -) c -INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n17 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: src3 - filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 110 Data size: 58740 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n17 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string), _col4 (type: int), _col5 (type: string) - outputColumnNames: c1, c2, c3, c4, c5, c6 - Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll'), compute_stats(c5, 'hll'), compute_stats(c6, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 2592 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n17 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3, c4, c5, c6 - Column Types: int, string, int, string, int, string - Table: default.dest1_n17 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 2592 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4), compute_stats(VALUE._col5) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 2640 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 2640 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - LEFT OUTER JOIN - ( - FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 - ) c - ON (a.c1 = c.c5) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 -) c -INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n17 -POSTHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - FULL OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - LEFT OUTER JOIN - ( - FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 - ) c - ON (a.c1 = c.c5) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 -) c -INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n17 -POSTHOOK: Lineage: dest1_n17.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n17.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n17.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n17.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n17.c5 EXPRESSION [(src)src3.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n17.c6 SIMPLE [(src)src3.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n17.* FROM dest1_n17 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n17 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n17.* FROM dest1_n17 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n17 -#### A masked pattern was here #### -11 val_11 NULL NULL NULL NULL -12 val_12 NULL NULL NULL NULL -12 val_12 NULL NULL NULL NULL -15 val_15 NULL NULL NULL NULL -15 val_15 NULL NULL NULL NULL -17 val_17 17 val_17 NULL NULL -18 val_18 18 val_18 NULL NULL -18 val_18 18 val_18 NULL NULL -18 val_18 18 val_18 NULL NULL -18 val_18 18 val_18 NULL NULL -19 val_19 19 val_19 NULL NULL -NULL NULL 20 val_20 NULL NULL -NULL NULL 24 val_24 NULL NULL -NULL NULL 24 val_24 NULL NULL diff --git ql/src/test/results/clientpositive/join8.q.out ql/src/test/results/clientpositive/join8.q.out deleted file mode 100644 index 4ebefb0163..0000000000 --- ql/src/test/results/clientpositive/join8.q.out +++ /dev/null @@ -1,220 +0,0 @@ -PREHOOK: query: CREATE TABLE dest1_n173(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest1_n173 -POSTHOOK: query: CREATE TABLE dest1_n173(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest1_n173 -PREHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n173 -POSTHOOK: query: EXPLAIN -FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n173 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col2 is null (type: boolean) - Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), null (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n173 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) - outputColumnNames: c1, c2, c3, c4 - Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n173 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: c1, c2, c3, c4 - Column Types: int, string, int, string - Table: default.dest1_n173 - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - -PREHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest1_n173 -POSTHOOK: query: FROM ( - FROM - ( - FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 - ) a - LEFT OUTER JOIN - ( - FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 - ) b - ON (a.c1 = b.c3) - SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 -) c -INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest1_n173 -POSTHOOK: Lineage: dest1_n173.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n173.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: dest1_n173.c3 EXPRESSION [] -POSTHOOK: Lineage: dest1_n173.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: SELECT dest1_n173.* FROM dest1_n173 -PREHOOK: type: QUERY -PREHOOK: Input: default@dest1_n173 -#### A masked pattern was here #### -POSTHOOK: query: SELECT dest1_n173.* FROM dest1_n173 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dest1_n173 -#### A masked pattern was here #### -11 val_11 NULL NULL -12 val_12 NULL NULL -12 val_12 NULL NULL -15 val_15 NULL NULL -15 val_15 NULL NULL diff --git ql/src/test/results/clientpositive/join_alt_syntax.q.out ql/src/test/results/clientpositive/join_alt_syntax.q.out deleted file mode 100644 index 1d30c5ac46..0000000000 --- ql/src/test/results/clientpositive/join_alt_syntax.q.out +++ /dev/null @@ -1,786 +0,0 @@ -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain select p1.p_name, p2.p_name -from part p1 , part p2 -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name -from part p1 , part p2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: p2 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1 - Statistics: Num rows: 676 Data size: 163592 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 676 Data size: 163592 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 ,part p2 ,part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 ,part p2 ,part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 , (select p_name from part) p2 ,part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 , (select p_name from part) p2 ,part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: part - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 , part p2 , part p3 -where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name -from part p1 , part p2 , part p3 -where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 702 Data size: 260442 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col0 + _col3) = _col3) (type: boolean) - Statistics: Num rows: 351 Data size: 130221 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 351 Data size: 127413 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 351 Data size: 127413 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name -from part p1 , part p2 join part p3 on p2.p_name = p1.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name -from part p1 , part p2 join part p3 on p2.p_name = p1.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: string) - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col1, _col3, _col5, _col6 - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name -from part p1 join part p2 on p2.p_name = p1.p_name , part p3 , part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name -from part p1 join part p2 on p2.p_name = p1.p_name , part p3 , part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: string) - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col1, _col3, _col5, _col6 - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_by_range_rule_not_null.q.out ql/src/test/results/clientpositive/join_by_range_rule_not_null.q.out deleted file mode 100644 index dcd59a953f..0000000000 --- ql/src/test/results/clientpositive/join_by_range_rule_not_null.q.out +++ /dev/null @@ -1,442 +0,0 @@ -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key < b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key < b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col0 < _col2) (type: boolean) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key AND a.value >= b.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key AND a.value >= b.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: b - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col1 >= _col3) (type: boolean) - Statistics: Num rows: 13 Data size: 4589 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 13 Data size: 4589 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: b - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col0 > _col3) (type: boolean) - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[6][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.key OR 1 = 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.key OR 1 = 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[4][tables = [a, b]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key IS DISTINCT FROM b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key IS DISTINCT FROM b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string), value (type: string) - TableScan - alias: b - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string), value (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col0 IS DISTINCT FROM _col5) (type: boolean) - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_1.q.out ql/src/test/results/clientpositive/join_cond_pushdown_1.q.out deleted file mode 100644 index 28fcdeeece..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_1.q.out +++ /dev/null @@ -1,489 +0,0 @@ -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name and p2.p_name = p3.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name and p2.p_name = p3.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 702 Data size: 1303614 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col0 + _col18) = _col18) (type: boolean) - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_partkey = 1 and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_partkey = 1 and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: ((p_partkey = 1) and p_name is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((p_partkey = 1) and p_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 619 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 - Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 - Statistics: Num rows: 26 Data size: 48178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_2.q.out ql/src/test/results/clientpositive/join_cond_pushdown_2.q.out deleted file mode 100644 index aac12821fc..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_2.q.out +++ /dev/null @@ -1,348 +0,0 @@ -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 on p2.p_name = p3.p_name and p1.p_name = p4.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 on p2.p_name = p3.p_name and p1.p_name = p4.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p4 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 on p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 on p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col9 (type: int), _col10 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col9 (type: int), _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col9 (type: int), _col10 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_3.q.out ql/src/test/results/clientpositive/join_cond_pushdown_3.q.out deleted file mode 100644 index d75d553cad..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_3.q.out +++ /dev/null @@ -1,497 +0,0 @@ -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p1.p_name = p2.p_name and p2.p_name = p3.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_name = p1.p_name and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_name = p1.p_name and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 702 Data size: 1303614 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((_col0 + _col18) = _col18) (type: boolean) - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_partkey = 1 and p3.p_name = p2.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 -where p2.p_partkey = 1 and p3.p_name = p2.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: ((p_partkey = 1) and p_name is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((p_partkey = 1) and p_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 619 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 - Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 - Statistics: Num rows: 26 Data size: 48178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_4.q.out ql/src/test/results/clientpositive/join_cond_pushdown_4.q.out deleted file mode 100644 index edb0c612c4..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_4.q.out +++ /dev/null @@ -1,352 +0,0 @@ -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_name = p4.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_name = p4.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p4 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 -where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2.p_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col9 (type: int), _col10 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col9 (type: int), _col10 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col9 (type: int), _col10 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_unqual1.q.out ql/src/test/results/clientpositive/join_cond_pushdown_unqual1.q.out deleted file mode 100644 index 18f828af2c..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_unqual1.q.out +++ /dev/null @@ -1,569 +0,0 @@ -PREHOOK: query: create table part2_n0( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part2_n0 -POSTHOOK: query: create table part2_n0( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part2_n0 -PREHOOK: query: create table part3_n0( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part3_n0 -POSTHOOK: query: create table part3_n0( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part3_n0 -PREHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p1.p_name = p2_name and p2_name = p3_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n0 -PREHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p1.p_name = p2_name and p2_name = p3_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n0 -POSTHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_name = p1.p_name and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n0 -PREHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_name = p1.p_name and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n0 -POSTHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey + p_partkey = p1.p_partkey and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n0 -PREHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey + p_partkey = p1.p_partkey and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n0 -POSTHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col0 + _col18) = _col18) (type: boolean) - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey = 1 and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n0 -PREHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey = 1 and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n0 -POSTHOOK: Input: default@part3_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: ((p2_partkey = 1) and p2_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((p2_partkey = 1) and p2_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_unqual2.q.out ql/src/test/results/clientpositive/join_cond_pushdown_unqual2.q.out deleted file mode 100644 index b491187560..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_unqual2.q.out +++ /dev/null @@ -1,412 +0,0 @@ -PREHOOK: query: create table part2( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part2 -POSTHOOK: query: create table part2( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part2 -PREHOOK: query: create table part3( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part3 -POSTHOOK: query: create table part3( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part3 -PREHOOK: query: explain select * -from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 on p2_name = p3_name and p1.p_name = p4.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2 -PREHOOK: Input: default@part3 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 on p2_name = p3_name and p1.p_name = p4.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2 -POSTHOOK: Input: default@part3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col19 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col19 (type: string) - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col18 (type: int), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) - TableScan - alias: p4 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col19 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 on p2_name = p3_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2 -PREHOOK: Input: default@part3 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 on p2_name = p3_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2 -POSTHOOK: Input: default@part3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: (p2_name is not null and p2_partkey is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (p2_name is not null and p2_partkey is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col18 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col18 (type: int) - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col18 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out deleted file mode 100644 index 16a13efcdc..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_unqual3.q.out +++ /dev/null @@ -1,577 +0,0 @@ -PREHOOK: query: create table part2_n5( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part2_n5 -POSTHOOK: query: create table part2_n5( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part2_n5 -PREHOOK: query: create table part3_n2( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part3_n2 -POSTHOOK: query: create table part3_n2( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part3_n2 -PREHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p1.p_name = p2_name and p2_name = p3_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n5 -PREHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p1.p_name = p2_name and p2_name = p3_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n5 -POSTHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_name = p1.p_name and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n5 -PREHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_name = p1.p_name and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n5 -POSTHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n5 -PREHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n5 -POSTHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col0 + _col18) = _col18) (type: boolean) - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -Warning: Shuffle Join JOIN[13][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Stage-2:MAPRED' is a cross product -PREHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_partkey = 1 and p3_name = p2_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n5 -PREHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n5 p2 join part3_n2 p3 -where p2_partkey = 1 and p3_name = p2_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n5 -POSTHOOK: Input: default@part3_n2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p2 - filterExpr: ((p2_partkey = 1) and p2_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((p2_partkey = 1) and p2_name is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) - TableScan - alias: p1 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out deleted file mode 100644 index 2a1f8a4ddc..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_unqual4.q.out +++ /dev/null @@ -1,416 +0,0 @@ -PREHOOK: query: create table part2_n4( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part2_n4 -POSTHOOK: query: create table part2_n4( - p2_partkey INT, - p2_name STRING, - p2_mfgr STRING, - p2_brand STRING, - p2_type STRING, - p2_size INT, - p2_container STRING, - p2_retailprice DOUBLE, - p2_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part2_n4 -PREHOOK: query: create table part3_n1( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@part3_n1 -POSTHOOK: query: create table part3_n1( - p3_partkey INT, - p3_name STRING, - p3_mfgr STRING, - p3_brand STRING, - p3_type STRING, - p3_size INT, - p3_container STRING, - p3_retailprice DOUBLE, - p3_comment STRING -) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@part3_n1 -PREHOOK: query: explain select * -from part p1 join part2_n4 p2 join part3_n1 p3 on p1.p_name = p2_name join part p4 -where p2_name = p3_name and p1.p_name = p4.p_name -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n4 -PREHOOK: Input: default@part3_n1 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n4 p2 join part3_n1 p3 on p1.p_name = p2_name join part p4 -where p2_name = p3_name and p1.p_name = p4.p_name -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n4 -POSTHOOK: Input: default@part3_n1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p2_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col19 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col19 (type: string) - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col18 (type: int), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) - TableScan - alias: p4 - filterExpr: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_name is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col19 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * -from part p1 join part2_n4 p2 join part3_n1 p3 on p2_name = p1.p_name join part p4 -where p2_name = p3_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2_partkey -PREHOOK: type: QUERY -PREHOOK: Input: default@part -PREHOOK: Input: default@part2_n4 -PREHOOK: Input: default@part3_n1 -#### A masked pattern was here #### -POSTHOOK: query: explain select * -from part p1 join part2_n4 p2 join part3_n1 p3 on p2_name = p1.p_name join part p4 -where p2_name = p3_name and p1.p_partkey = p4.p_partkey - and p1.p_partkey = p2_partkey -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -POSTHOOK: Input: default@part2_n4 -POSTHOOK: Input: default@part3_n1 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: (p2_name is not null and p2_partkey is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (p2_name is not null and p2_partkey is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - TableScan - alias: p3 - filterExpr: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: p3_name is not null (type: boolean) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) - TableScan - alias: p1 - filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_name is not null and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: string) - 1 _col0 (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col18 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col18 (type: int) - Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) - TableScan - alias: p4 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col18 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out deleted file mode 100644 index 202ef40508..0000000000 --- ql/src/test/results/clientpositive/join_cond_pushdown_unqual5.q.out +++ /dev/null @@ -1,215 +0,0 @@ -PREHOOK: query: CREATE TABLE ltable (index int, la int, lk1 string, lk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@ltable -POSTHOOK: query: CREATE TABLE ltable (index int, la int, lk1 string, lk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@ltable -PREHOOK: query: CREATE TABLE rtable (ra int, rk1 string, rk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@rtable -POSTHOOK: query: CREATE TABLE rtable (ra int, rk1 string, rk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@rtable -PREHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@ltable -POSTHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@ltable -POSTHOOK: Lineage: ltable.index SCRIPT [] -POSTHOOK: Lineage: ltable.la EXPRESSION [] -POSTHOOK: Lineage: ltable.lk1 SCRIPT [] -POSTHOOK: Lineage: ltable.lk2 SCRIPT [] -PREHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@rtable -POSTHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072') -POSTHOOK: type: QUERY -POSTHOOK: Input: _dummy_database@_dummy_table -POSTHOOK: Output: default@rtable -POSTHOOK: Lineage: rtable.ra SCRIPT [] -POSTHOOK: Lineage: rtable.rk1 SCRIPT [] -POSTHOOK: Lineage: rtable.rk2 SCRIPT [] -PREHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -PREHOOK: type: QUERY -PREHOOK: Input: default@ltable -PREHOOK: Input: default@rtable -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ltable -POSTHOOK: Input: default@rtable -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: l - Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: index (type: int), la (type: int), lk1 (type: string), lk2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: string), _col3 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col2 (type: string), _col3 (type: string) - Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: int) - TableScan - alias: r - filterExpr: (rk1 is not null and rk2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (rk1 is not null and rk2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ra (type: int), rk1 (type: string), rk2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string), _col2 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col2 (type: string), _col3 (type: string) - 1 _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 3 Data size: 1160 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (CASE WHEN (_col1 is not null) THEN (CAST( _col1 AS STRING)) ELSE ('EMPTY') END = CASE WHEN (_col4 is not null) THEN (CAST( _col4 AS STRING)) ELSE ('EMPTY') END) (type: boolean) - Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -PREHOOK: type: QUERY -PREHOOK: Input: default@ltable -PREHOOK: Input: default@rtable -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ltable -POSTHOOK: Input: default@rtable -#### A masked pattern was here #### -2 NULL CD5415192225530 00071 NULL NULL NULL -PREHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -PREHOOK: type: QUERY -PREHOOK: Input: default@ltable -PREHOOK: Input: default@rtable -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ltable -POSTHOOK: Input: default@rtable -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:r - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:r - TableScan - alias: r - filterExpr: (rk1 is not null and rk2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (rk1 is not null and rk2 is not null) (type: boolean) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ra (type: int), rk1 (type: string), rk2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col2 (type: string), _col3 (type: string) - 1 _col1 (type: string), _col2 (type: string) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: l - Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: index (type: int), la (type: int), lk1 (type: string), lk2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col2 (type: string), _col3 (type: string) - 1 _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 3 Data size: 1160 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (CASE WHEN (_col1 is not null) THEN (CAST( _col1 AS STRING)) ELSE ('EMPTY') END = CASE WHEN (_col4 is not null) THEN (CAST( _col4 AS STRING)) ELSE ('EMPTY') END) (type: boolean) - Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -PREHOOK: type: QUERY -PREHOOK: Input: default@ltable -PREHOOK: Input: default@rtable -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') -POSTHOOK: type: QUERY -POSTHOOK: Input: default@ltable -POSTHOOK: Input: default@rtable -#### A masked pattern was here #### -2 NULL CD5415192225530 00071 NULL NULL NULL diff --git ql/src/test/results/clientpositive/join_filters_overlap.q.out ql/src/test/results/clientpositive/join_filters_overlap.q.out deleted file mode 100644 index 90f441c421..0000000000 --- ql/src/test/results/clientpositive/join_filters_overlap.q.out +++ /dev/null @@ -1,2033 +0,0 @@ -PREHOOK: query: create table a_n4 as SELECT 100 as key, a_n4.value as value FROM src LATERAL VIEW explode(array(40, 50, 60)) a_n4 as value limit 3 -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@a_n4 -POSTHOOK: query: create table a_n4 as SELECT 100 as key, a_n4.value as value FROM src LATERAL VIEW explode(array(40, 50, 60)) a_n4 as value limit 3 -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@a_n4 -POSTHOOK: Lineage: a_n4.key SIMPLE [] -POSTHOOK: Lineage: a_n4.value SCRIPT [] -PREHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` -FROM (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` -FROM `default`.`a_n4`) AS `t` -LEFT JOIN (SELECT `key`, CAST(50 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 50 AND `key` IS NOT NULL) AS `t1` ON `t`.`key` = `t1`.`key` AND `t`.`=3` -LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t`.`key` = `t3`.`key` AND `t`.`=` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n4 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) - auto parallelism: false - TableScan - alias: b - filterExpr: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 50 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_0:a_n4, $hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col2} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col4, _col5 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5 - columns.types int,int,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col4 (type: int), _col5 (type: int) - auto parallelism: false - TableScan - alias: c - filterExpr: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 60 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5 - columns.types int,int,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5 - columns.types int,int,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_2:c] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col1} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4, _col5, _col6, _col7 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5 - columns.types int:int:int:int:int:int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 40 NULL NULL NULL NULL -100 50 100 50 NULL NULL -100 60 NULL NULL 100 60 -PREHOOK: query: select /*+ MAPJOIN(b,c)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ MAPJOIN(b,c)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 40 NULL NULL NULL NULL -100 50 100 50 NULL NULL -100 60 NULL NULL 100 60 -PREHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` -FROM (SELECT `key`, CAST(50 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 50 AND `key` IS NOT NULL) AS `t0` -RIGHT JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` -FROM `default`.`a_n4`) AS `t1` ON `t0`.`key` = `t1`.`key` AND `t1`.`=3` -LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t1`.`key` = `t3`.`key` AND `t1`.`=` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n4 - filterExpr: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 50 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int) - auto parallelism: false - TableScan - alias: b - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_0:a_n4, $hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter mappings: - 1 [0, 1] - filter predicates: - 0 - 1 {VALUE._col2} - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col2 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col2 (type: int) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: boolean) - auto parallelism: false - TableScan - alias: c - filterExpr: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 60 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_2:c] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col3} - 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int), _col6 (type: int), _col7 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5 - columns.types int:int:int:int:int:int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 50 100 50 NULL NULL -NULL NULL 100 40 NULL NULL -NULL NULL 100 60 100 60 -PREHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 50 100 50 NULL NULL -NULL NULL 100 40 NULL NULL -NULL NULL 100 60 100 60 -PREHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` -FROM (SELECT `key`, CAST(50 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 50 AND `key` IS NOT NULL) AS `t0` -RIGHT JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` -FROM `default`.`a_n4`) AS `t1` ON `t0`.`key` = `t1`.`key` AND `t1`.`=3` -LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t1`.`key` = `t3`.`key` AND `t1`.`=` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n4 - filterExpr: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 50 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int) - auto parallelism: false - TableScan - alias: b - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_0:a_n4, $hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter mappings: - 1 [0, 1] - filter predicates: - 0 - 1 {VALUE._col2} - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col2 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col2 (type: int) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: boolean) - auto parallelism: false - TableScan - alias: c - filterExpr: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 60 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col4 - columns.types int,int,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_2:c] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col3} - 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int), _col6 (type: int), _col7 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5 - columns.types int:int:int:int:int:int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 50 100 50 NULL NULL -NULL NULL 100 40 NULL NULL -NULL NULL 100 60 100 60 -PREHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 50 100 50 NULL NULL -NULL NULL 100 40 NULL NULL -NULL NULL 100 60 100 60 -PREHOOK: query: explain extended select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain extended select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t0`.`key` AS `key1`, `t0`.`value` AS `value1`, `t2`.`key` AS `key2`, `t2`.`value` AS `value2`, `t4`.`key` AS `key3`, `t4`.`value` AS `value3` -FROM (SELECT `key`, `value`, `value` = 40 AS `=`, `value` = 50 AS `=3` -FROM `default`.`a_n4`) AS `t` -FULL JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` -FROM `default`.`a_n4`) AS `t0` ON `t`.`key` = `t0`.`key` AND `t`.`=3` AND `t0`.`=3` -LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 60 AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = `t2`.`key` AND `t0`.`=` -LEFT JOIN (SELECT `key`, CAST(40 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 40 AND `key` IS NOT NULL) AS `t4` ON `t`.`key` = `t4`.`key` AND `t`.`=` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n4 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 40) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) - auto parallelism: false - TableScan - alias: b - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_0:a_n4, $hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - 1 [0, 1] - filter predicates: - 0 {VALUE._col2} - 1 {VALUE._col2} - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col6 - Statistics: Num rows: 9 Data size: 216 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col6 - columns.types int,int,boolean,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col4 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col4 (type: int) - Statistics: Num rows: 9 Data size: 216 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: boolean), _col5 (type: int), _col6 (type: boolean) - auto parallelism: false - TableScan - alias: c - filterExpr: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 60 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col6 - columns.types int,int,boolean,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col6 - columns.types int,int,boolean,int,int,boolean - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_2:c] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col5} - 1 - keys: - 0 _col4 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col8, _col9 - Statistics: Num rows: 9 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col8,_col9 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 9 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col4 (type: int), _col5 (type: int), _col8 (type: int), _col9 (type: int) - auto parallelism: false - TableScan - alias: d - filterExpr: ((value = 40) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 40) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 40 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10005 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col8,_col9 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col4,_col5,_col8,_col9 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_3:d] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col1} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9, _col10, _col11 - Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col5 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: int), _col11 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 - columns.types int:int:int:int:int:int:int:int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 40 NULL NULL NULL NULL 100 40 -100 50 100 50 NULL NULL NULL NULL -100 60 NULL NULL NULL NULL NULL NULL -NULL NULL 100 40 NULL NULL NULL NULL -NULL NULL 100 60 100 60 NULL NULL -PREHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2`, `t5`.`key` AS `key3`, `t5`.`value` AS `value3` -FROM (SELECT `key`, `value`, `value` = 40 AS `=`, `value` = 60 AS `=3`, `value` = 50 AS `=4` -FROM `default`.`a_n4`) AS `t` -LEFT JOIN (SELECT `key`, CAST(50 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 50 AND `key` IS NOT NULL) AS `t1` ON `t`.`key` = `t1`.`key` AND `t`.`=4` -LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t`.`key` = `t3`.`key` AND `t`.`=3` -LEFT JOIN (SELECT `key`, CAST(40 AS INTEGER) AS `value` -FROM `default`.`a_n4` -WHERE `value` = 40 AND `key` IS NOT NULL) AS `t5` ON `t`.`key` = `t5`.`key` AND `t`.`=` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n4 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: int), value (type: int), (value = 40) (type: boolean), (value = 60) (type: boolean), (value = 50) (type: boolean) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean), _col4 (type: boolean) - auto parallelism: false - TableScan - alias: b - filterExpr: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 50) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 50 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_0:a_n4, $hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col3} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col5, _col6 - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col5,_col6 - columns.types int,int,boolean,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean), _col5 (type: int), _col6 (type: int) - auto parallelism: false - TableScan - alias: c - filterExpr: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 60) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 60 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col5,_col6 - columns.types int,int,boolean,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3,_col5,_col6 - columns.types int,int,boolean,boolean,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_2:c] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col2} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col5, _col6, _col7, _col8 - Statistics: Num rows: 3 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col5,_col6,_col7,_col8 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 3 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: int), _col2 (type: boolean), _col5 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int) - auto parallelism: false - TableScan - alias: d - filterExpr: ((value = 40) and key is not null) (type: boolean) - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((value = 40) and key is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), 40 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: int) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: int) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10005 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col5,_col6,_col7,_col8 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col5,_col6,_col7,_col8 - columns.types int,int,boolean,int,int,int,int - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe -#### A masked pattern was here #### - Partition - base file name: a_n4 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types int:int -#### A masked pattern was here #### - name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.a_n4 - name: default.a_n4 - Truncated Path -> Alias: - /a_n4 [$hdt$_3:d] -#### A masked pattern was here #### - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter mappings: - 0 [1, 1] - filter predicates: - 0 {VALUE._col1} - 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8, _col9, _col10 - Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 - columns.types int:int:int:int:int:int:int:int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 40 NULL NULL NULL NULL 100 40 -100 50 100 50 NULL NULL NULL NULL -100 60 NULL NULL 100 60 NULL NULL -PREHOOK: query: select /*+ MAPJOIN(b,c, d)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n4 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ MAPJOIN(b,c, d)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n4 -#### A masked pattern was here #### -100 40 NULL NULL NULL NULL 100 40 -100 50 100 50 NULL NULL NULL NULL -100 60 NULL NULL 100 60 NULL NULL diff --git ql/src/test/results/clientpositive/join_grp_diff_keys.q.out ql/src/test/results/clientpositive/join_grp_diff_keys.q.out deleted file mode 100644 index bc954d2cec..0000000000 --- ql/src/test/results/clientpositive/join_grp_diff_keys.q.out +++ /dev/null @@ -1,277 +0,0 @@ -PREHOOK: query: create table split (id int, line_id int, orders string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@split -POSTHOOK: query: create table split (id int, line_id int, orders string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@split -PREHOOK: query: create table bar (id int, line_id int, orders string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@bar -POSTHOOK: query: create table bar (id int, line_id int, orders string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@bar -PREHOOK: query: create table foo (id int, line_id int, orders string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@foo -POSTHOOK: query: create table foo (id int, line_id int, orders string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@foo -PREHOOK: query: create table forecast (id int, line_id int, orders string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@forecast -POSTHOOK: query: create table forecast (id int, line_id int, orders string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@forecast -PREHOOK: query: explain -SELECT foo.id, count(*) as factor from - foo JOIN bar ON (foo.id = bar.id and foo.line_id = bar.line_id) - JOIN split ON (foo.id = split.id and foo.line_id = split.line_id) - JOIN forecast ON (foo.id = forecast.id AND foo.line_id = forecast.line_id) - WHERE foo.orders != 'blah' - group by foo.id -PREHOOK: type: QUERY -PREHOOK: Input: default@bar -PREHOOK: Input: default@foo -PREHOOK: Input: default@forecast -PREHOOK: Input: default@split -#### A masked pattern was here #### -POSTHOOK: query: explain -SELECT foo.id, count(*) as factor from - foo JOIN bar ON (foo.id = bar.id and foo.line_id = bar.line_id) - JOIN split ON (foo.id = split.id and foo.line_id = split.line_id) - JOIN forecast ON (foo.id = forecast.id AND foo.line_id = forecast.line_id) - WHERE foo.orders != 'blah' - group by foo.id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@bar -POSTHOOK: Input: default@foo -POSTHOOK: Input: default@forecast -POSTHOOK: Input: default@split -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-3 - Stage-0 depends on stages: Stage-4 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: foo - filterExpr: ((orders <> 'blah') and id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((orders <> 'blah') and id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), line_id (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: bar - filterExpr: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), line_id (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: int) - 1 _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: split - filterExpr: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), line_id (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: int) - 1 _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: forecast - filterExpr: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (id is not null and line_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), line_id (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int), _col1 (type: int) - 1 _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col0 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: drop table split -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@split -PREHOOK: Output: default@split -POSTHOOK: query: drop table split -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@split -POSTHOOK: Output: default@split -PREHOOK: query: drop table bar -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@bar -PREHOOK: Output: default@bar -POSTHOOK: query: drop table bar -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@bar -POSTHOOK: Output: default@bar -PREHOOK: query: drop table foo -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@foo -PREHOOK: Output: default@foo -POSTHOOK: query: drop table foo -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@foo -POSTHOOK: Output: default@foo -PREHOOK: query: drop table forecast -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@forecast -PREHOOK: Output: default@forecast -POSTHOOK: query: drop table forecast -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@forecast -POSTHOOK: Output: default@forecast diff --git ql/src/test/results/clientpositive/join_hive_626.q.out ql/src/test/results/clientpositive/join_hive_626.q.out deleted file mode 100644 index cf0e61c87c..0000000000 --- ql/src/test/results/clientpositive/join_hive_626.q.out +++ /dev/null @@ -1,206 +0,0 @@ -PREHOOK: query: create table hive_foo (foo_id int, foo_name string, foo_a string, foo_b string, -foo_c string, foo_d string) row format delimited fields terminated by ',' -stored as textfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@hive_foo -POSTHOOK: query: create table hive_foo (foo_id int, foo_name string, foo_a string, foo_b string, -foo_c string, foo_d string) row format delimited fields terminated by ',' -stored as textfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@hive_foo -PREHOOK: query: create table hive_bar (bar_id int, bar_0 int, foo_id int, bar_1 int, bar_name -string, bar_a string, bar_b string, bar_c string, bar_d string) row format -delimited fields terminated by ',' stored as textfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@hive_bar -POSTHOOK: query: create table hive_bar (bar_id int, bar_0 int, foo_id int, bar_1 int, bar_name -string, bar_a string, bar_b string, bar_c string, bar_d string) row format -delimited fields terminated by ',' stored as textfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@hive_bar -PREHOOK: query: create table hive_count (bar_id int, n int) row format delimited fields -terminated by ',' stored as textfile -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@hive_count -POSTHOOK: query: create table hive_count (bar_id int, n int) row format delimited fields -terminated by ',' stored as textfile -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@hive_count -PREHOOK: query: load data local inpath '../../data/files/hive_626_foo.txt' overwrite into table hive_foo -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@hive_foo -POSTHOOK: query: load data local inpath '../../data/files/hive_626_foo.txt' overwrite into table hive_foo -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@hive_foo -PREHOOK: query: load data local inpath '../../data/files/hive_626_bar.txt' overwrite into table hive_bar -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@hive_bar -POSTHOOK: query: load data local inpath '../../data/files/hive_626_bar.txt' overwrite into table hive_bar -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@hive_bar -PREHOOK: query: load data local inpath '../../data/files/hive_626_count.txt' overwrite into table hive_count -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@hive_count -POSTHOOK: query: load data local inpath '../../data/files/hive_626_count.txt' overwrite into table hive_count -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@hive_count -PREHOOK: query: explain -select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = -hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id -PREHOOK: type: QUERY -PREHOOK: Input: default@hive_bar -PREHOOK: Input: default@hive_count -PREHOOK: Input: default@hive_foo -#### A masked pattern was here #### -POSTHOOK: query: explain -select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = -hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hive_bar -POSTHOOK: Input: default@hive_count -POSTHOOK: Input: default@hive_foo -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: hive_bar - filterExpr: (foo_id is not null and bar_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (foo_id is not null and bar_id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: bar_id (type: int), foo_id (type: int), bar_name (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: string) - TableScan - alias: hive_foo - filterExpr: foo_id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: foo_id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: foo_id (type: int), foo_name (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col2, _col4 - Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE - value expressions: _col2 (type: string), _col4 (type: string) - TableScan - alias: hive_count - filterExpr: bar_id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: bar_id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: bar_id (type: int), n (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col2, _col4, _col6 - Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col4 (type: string), _col2 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = -hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id -PREHOOK: type: QUERY -PREHOOK: Input: default@hive_bar -PREHOOK: Input: default@hive_count -PREHOOK: Input: default@hive_foo -#### A masked pattern was here #### -POSTHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = -hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@hive_bar -POSTHOOK: Input: default@hive_count -POSTHOOK: Input: default@hive_foo -#### A masked pattern was here #### -foo1 bar10 2 diff --git ql/src/test/results/clientpositive/join_merge_multi_expressions.q.out ql/src/test/results/clientpositive/join_merge_multi_expressions.q.out deleted file mode 100644 index 28a1238930..0000000000 --- ql/src/test/results/clientpositive/join_merge_multi_expressions.q.out +++ /dev/null @@ -1,173 +0,0 @@ -PREHOOK: query: explain -select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain -select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: (key is not null and hr is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), hr (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: b - filterExpr: (key is not null and hr is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), hr (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col1 (type: string) - 1 _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12658 Data size: 3430318 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 12658 Data size: 3430318 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: c - filterExpr: (hr is not null and key is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), hr (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col1 (type: string) - 1 _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 80113 Data size: 640904 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -42464 diff --git ql/src/test/results/clientpositive/join_merging.q.out ql/src/test/results/clientpositive/join_merging.q.out deleted file mode 100644 index ec39424971..0000000000 --- ql/src/test/results/clientpositive/join_merging.q.out +++ /dev/null @@ -1,263 +0,0 @@ -PREHOOK: query: explain select p1.p_size, p2.p_size -from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey - right outer join part p3 on p2.p_partkey = p3.p_partkey and - p1.p_size > 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_size, p2.p_size -from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey - right outer join part p3 on p2.p_partkey = p3.p_partkey and - p1.p_size > 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: p_partkey is not null (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_size (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: p1 - filterExpr: ((p_size > 10) and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((p_size > 10) and p_partkey is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_size (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col3 (type: int), _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col1 (type: int) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: int), _col3 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select p1.p_size, p2.p_size -from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey - right outer join part p3 on p2.p_partkey = p3.p_partkey and - p1.p_size > 10 and p1.p_size > p2.p_size + 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@part -#### A masked pattern was here #### -POSTHOOK: query: explain select p1.p_size, p2.p_size -from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey - right outer join part p3 on p2.p_partkey = p3.p_partkey and - p1.p_size > 10 and p1.p_size > p2.p_size + 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@part -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: (p_partkey is not null and p_size is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (p_partkey is not null and p_size is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_size (type: int), (p_size + 10) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 26 Data size: 312 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 312 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int), _col2 (type: int) - TableScan - alias: p1 - filterExpr: ((p_size > 10) and p_partkey is not null) (type: boolean) - Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((p_size > 10) and p_partkey is not null) (type: boolean) - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int), p_size (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col4 - Statistics: Num rows: 21 Data size: 336 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col4 > _col2) (type: boolean) - Statistics: Num rows: 7 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col0 (type: int), _col1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 7 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p3 - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: p_partkey (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 7 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col2 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col1 (type: int) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: int), _col3 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_on_varchar.q.out ql/src/test/results/clientpositive/join_on_varchar.q.out deleted file mode 100644 index 01bd1e3ae9..0000000000 --- ql/src/test/results/clientpositive/join_on_varchar.q.out +++ /dev/null @@ -1,156 +0,0 @@ -PREHOOK: query: create table tbl1_n3(c1 varchar(10), intcol int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl1_n3 -POSTHOOK: query: create table tbl1_n3(c1 varchar(10), intcol int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl1_n3 -PREHOOK: query: create table tbl2_n2(c2 varchar(30)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@tbl2_n2 -POSTHOOK: query: create table tbl2_n2(c2 varchar(30)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@tbl2_n2 -PREHOOK: query: insert into table tbl1_n3 select repeat('t', 10), 11 from src limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl1_n3 -POSTHOOK: query: insert into table tbl1_n3 select repeat('t', 10), 11 from src limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl1_n3 -POSTHOOK: Lineage: tbl1_n3.c1 EXPRESSION [] -POSTHOOK: Lineage: tbl1_n3.intcol SIMPLE [] -PREHOOK: query: insert into table tbl1_n3 select repeat('s', 10), 22 from src limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl1_n3 -POSTHOOK: query: insert into table tbl1_n3 select repeat('s', 10), 22 from src limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl1_n3 -POSTHOOK: Lineage: tbl1_n3.c1 EXPRESSION [] -POSTHOOK: Lineage: tbl1_n3.intcol SIMPLE [] -PREHOOK: query: insert into table tbl2_n2 select concat(repeat('t', 10), 'ppp') from src limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl2_n2 -POSTHOOK: query: insert into table tbl2_n2 select concat(repeat('t', 10), 'ppp') from src limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl2_n2 -POSTHOOK: Lineage: tbl2_n2.c2 EXPRESSION [] -PREHOOK: query: insert into table tbl2_n2 select repeat('s', 10) from src limit 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@tbl2_n2 -POSTHOOK: query: insert into table tbl2_n2 select repeat('s', 10) from src limit 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@tbl2_n2 -POSTHOOK: Lineage: tbl2_n2.c2 EXPRESSION [] -PREHOOK: query: explain -select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 -PREHOOK: type: QUERY -PREHOOK: Input: default@tbl1_n3 -PREHOOK: Input: default@tbl2_n2 -#### A masked pattern was here #### -POSTHOOK: query: explain -select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tbl1_n3 -POSTHOOK: Input: default@tbl2_n2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-5 is a root stage - Stage-2 depends on stages: Stage-5 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-5 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:tbl2_n2 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:tbl2_n2 - TableScan - alias: tbl2_n2 - filterExpr: c2 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: c2 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: c2 (type: varchar(30)) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: varchar(30)) - 1 _col0 (type: varchar(30)) - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: tbl1_n3 - filterExpr: c1 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: c1 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: c1 (type: varchar(10)) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: varchar(30)) - 1 _col0 (type: varchar(30)) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: varchar(10)), _col1 (type: varchar(30)) - null sort order: zz - sort order: ++ - Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Local Work: - Map Reduce Local Work - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: varchar(10)), KEY.reducesinkkey1 (type: varchar(30)) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 -PREHOOK: type: QUERY -PREHOOK: Input: default@tbl1_n3 -PREHOOK: Input: default@tbl2_n2 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@tbl1_n3 -POSTHOOK: Input: default@tbl2_n2 -#### A masked pattern was here #### -ssssssssss ssssssssss diff --git ql/src/test/results/clientpositive/join_parse.q.out ql/src/test/results/clientpositive/join_parse.q.out deleted file mode 100644 index f37b58b4f5..0000000000 --- ql/src/test/results/clientpositive/join_parse.q.out +++ /dev/null @@ -1,603 +0,0 @@ -PREHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key)) -inner join src src1 on src1.value =srcpart.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key)) -inner join src src1 on src1.value =srcpart.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: src1 - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -select srcpart.key, src1.value from -(srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain -select srcpart.key, src1.value from -(srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: src1 - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value) -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: src1 - filterExpr: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: value is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value) -inner join src src2 on src2.key = src1.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain -select srcpart.key, src1.value from -((srcpart inner join src on srcpart.key = src.key) -inner join src src1 on src1.value =srcpart.value) -inner join src src2 on src2.key = src1.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-4 - Stage-4 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and value is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 791 Data size: 71981 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col4 - Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col4 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (value is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (value is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1 - Statistics: Num rows: 791 Data size: 71981 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/join_reorder.q.out ql/src/test/results/clientpositive/join_reorder.q.out deleted file mode 100644 index 14557a5b99..0000000000 --- ql/src/test/results/clientpositive/join_reorder.q.out +++ /dev/null @@ -1,724 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n37(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n37 -POSTHOOK: query: CREATE TABLE T1_n37(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n37 -PREHOOK: query: CREATE TABLE T2_n24(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n24 -POSTHOOK: query: CREATE TABLE T2_n24(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n24 -PREHOOK: query: CREATE TABLE T3_n8(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n8 -POSTHOOK: query: CREATE TABLE T3_n8(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n8 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n37 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n37 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n37 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n37 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n24 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t2_n24 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n24 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t2_n24 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n8 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t3_n8 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n8 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t3_n8 -PREHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT a.key, a.val, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT a.key, a.val, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(key) (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string), val (type: string) - TableScan - alias: c - filterExpr: (key + 1) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key + 1) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: (key + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (key + 1) (type: double) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double) - 1 (key + 1) (type: double) - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(key) (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string), val (type: string) - TableScan - alias: c - filterExpr: (key + 1) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key + 1) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: (key + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (key + 1) (type: double) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double) - 1 (key + 1) (type: double) - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT a.key, a.val, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -POSTHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT a.key, a.val, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -1 11 0 -1 11 0 -1 11 0 -3 13 2 -PREHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -POSTHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key -SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@t1_n37 -#### A masked pattern was here #### -1 11 0 -1 11 0 -1 11 0 -3 13 2 -PREHOOK: query: EXPLAIN FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT a.key, b.key, a.val, c.val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT a.key, b.key, a.val, c.val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col5 (type: string) - TableScan - alias: c - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: val (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 val (type: string) - outputColumnNames: _col0, _col1, _col5, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col5 (type: string), _col1 (type: string), _col11 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col5 (type: string) - TableScan - alias: c - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: val (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 val (type: string) - outputColumnNames: _col0, _col1, _col5, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col5 (type: string), _col1 (type: string), _col11 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT a.key, b.key, a.val, c.val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT a.key, b.key, a.val, c.val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -2 2 12 12 -7 NULL 17 17 -NULL NULL NULL 14 -NULL NULL NULL 16 -PREHOOK: query: FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: FROM T1_n37 a - LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) - RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) -SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -2 2 12 12 -7 NULL 17 17 -NULL NULL NULL 14 -NULL NULL NULL 16 -PREHOOK: query: EXPLAIN FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT a.key, b.key, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT a.key, b.key, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: c - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Unique Join 0 to 0 - Unique Join 0 to 0 - Unique Join 0 to 0 - keys: - 0 key (type: string), val (type: string) - 1 key (type: string), val (type: string) - 2 key (type: string), val (type: string) - outputColumnNames: _col0, _col5, _col10 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col5 (type: string), _col10 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: b - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: c - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string), val (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: key (type: string), val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Unique Join 0 to 0 - Unique Join 0 to 0 - Unique Join 0 to 0 - keys: - 0 key (type: string), val (type: string) - 1 key (type: string), val (type: string) - 2 key (type: string), val (type: string) - outputColumnNames: _col0, _col5, _col10 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col5 (type: string), _col10 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT a.key, b.key, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT a.key, b.key, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -1 NULL NULL -2 NULL 2 -3 3 NULL -7 NULL 7 -8 8 NULL -8 8 NULL -8 NULL NULL -NULL 2 NULL -NULL 4 4 -NULL 5 NULL -NULL NULL 6 -PREHOOK: query: FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n37 -PREHOOK: Input: default@t2_n24 -PREHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -POSTHOOK: query: FROM UNIQUEJOIN - PRESERVE T1_n37 a (a.key, a.val), - PRESERVE T2_n24 b (b.key, b.val), - PRESERVE T3_n8 c (c.key, c.val) -SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n37 -POSTHOOK: Input: default@t2_n24 -POSTHOOK: Input: default@t3_n8 -#### A masked pattern was here #### -1 NULL NULL -2 NULL 2 -3 3 NULL -7 NULL 7 -8 8 NULL -8 8 NULL -8 NULL NULL -NULL 2 NULL -NULL 4 4 -NULL 5 NULL -NULL NULL 6 diff --git ql/src/test/results/clientpositive/join_reorder2.q.out ql/src/test/results/clientpositive/join_reorder2.q.out deleted file mode 100644 index 76fdd600a1..0000000000 --- ql/src/test/results/clientpositive/join_reorder2.q.out +++ /dev/null @@ -1,444 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n49(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n49 -POSTHOOK: query: CREATE TABLE T1_n49(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n49 -PREHOOK: query: CREATE TABLE T2_n30(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n30 -POSTHOOK: query: CREATE TABLE T2_n30(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n30 -PREHOOK: query: CREATE TABLE T3_n10(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n10 -POSTHOOK: query: CREATE TABLE T3_n10(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n10 -PREHOOK: query: CREATE TABLE T4_n3(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T4_n3 -POSTHOOK: query: CREATE TABLE T4_n3(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T4_n3 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n49 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n49 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n49 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n49 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n30 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t2_n30 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n30 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t2_n30 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n10 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t3_n10 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n10 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t3_n10 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n3 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t4_n3 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n3 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t4_n3 -PREHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON b.key = c.key - JOIN T4_n3 d ON c.key = d.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n49 -PREHOOK: Input: default@t2_n30 -PREHOOK: Input: default@t3_n10 -PREHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON b.key = c.key - JOIN T4_n3 d ON c.key = d.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n49 -POSTHOOK: Input: default@t2_n30 -POSTHOOK: Input: default@t3_n10 -POSTHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col5 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col5 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col5 (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col11 (type: string) - TableScan - alias: d - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON b.key = c.key - JOIN T4_n3 d ON c.key = d.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n49 -PREHOOK: Input: default@t2_n30 -PREHOOK: Input: default@t3_n10 -PREHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON b.key = c.key - JOIN T4_n3 d ON c.key = d.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n49 -POSTHOOK: Input: default@t2_n30 -POSTHOOK: Input: default@t3_n10 -POSTHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -2 12 2 22 2 12 2 12 -PREHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON a.val = c.val - JOIN T4_n3 d ON a.key + 1 = d.key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n49 -PREHOOK: Input: default@t2_n30 -PREHOOK: Input: default@t3_n10 -PREHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON a.val = c.val - JOIN T4_n3 d ON a.key + 1 = d.key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n49 -POSTHOOK: Input: default@t2_n30 -POSTHOOK: Input: default@t3_n10 -POSTHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: (key is not null and val is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key is not null and val is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - filterExpr: (key is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col5 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: val is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 val (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: (_col0 + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (_col0 + 1) (type: double) - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) - TableScan - alias: d - filterExpr: (key + 1) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key + 1) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: (key + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (key + 1) (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string), val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 (_col0 + 1) (type: double) - 1 (key + 1) (type: double) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON a.val = c.val - JOIN T4_n3 d ON a.key + 1 = d.key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n49 -PREHOOK: Input: default@t2_n30 -PREHOOK: Input: default@t3_n10 -PREHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ * -FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key - JOIN T3_n10 c ON a.val = c.val - JOIN T4_n3 d ON a.key + 1 = d.key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n49 -POSTHOOK: Input: default@t2_n30 -POSTHOOK: Input: default@t3_n10 -POSTHOOK: Input: default@t4_n3 -#### A masked pattern was here #### -2 12 2 22 2 12 2 12 diff --git ql/src/test/results/clientpositive/join_reorder3.q.out ql/src/test/results/clientpositive/join_reorder3.q.out deleted file mode 100644 index 35ec179029..0000000000 --- ql/src/test/results/clientpositive/join_reorder3.q.out +++ /dev/null @@ -1,444 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n92(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n92 -POSTHOOK: query: CREATE TABLE T1_n92(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n92 -PREHOOK: query: CREATE TABLE T2_n57(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n57 -POSTHOOK: query: CREATE TABLE T2_n57(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n57 -PREHOOK: query: CREATE TABLE T3_n21(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n21 -POSTHOOK: query: CREATE TABLE T3_n21(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n21 -PREHOOK: query: CREATE TABLE T4_n10(key STRING, val STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T4_n10 -POSTHOOK: query: CREATE TABLE T4_n10(key STRING, val STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T4_n10 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n92 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n92 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n92 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n92 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n57 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t2_n57 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n57 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t2_n57 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n21 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t3_n21 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n21 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t3_n21 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n10 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t4_n10 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n10 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t4_n10 -PREHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON b.key = c.key - JOIN T4_n10 d ON c.key = d.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n92 -PREHOOK: Input: default@t2_n57 -PREHOOK: Input: default@t3_n21 -PREHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON b.key = c.key - JOIN T4_n10 d ON c.key = d.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n92 -POSTHOOK: Input: default@t2_n57 -POSTHOOK: Input: default@t3_n21 -POSTHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col5 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col5 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col5 (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col10 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col10 (type: string) - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col11 (type: string) - TableScan - alias: d - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col10 (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON b.key = c.key - JOIN T4_n10 d ON c.key = d.key -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n92 -PREHOOK: Input: default@t2_n57 -PREHOOK: Input: default@t3_n21 -PREHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON b.key = c.key - JOIN T4_n10 d ON c.key = d.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n92 -POSTHOOK: Input: default@t2_n57 -POSTHOOK: Input: default@t3_n21 -POSTHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -2 12 2 22 2 12 2 12 -PREHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON a.val = c.val - JOIN T4_n10 d ON a.key + 1 = d.key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n92 -PREHOOK: Input: default@t2_n57 -PREHOOK: Input: default@t3_n21 -PREHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON a.val = c.val - JOIN T4_n10 d ON a.key + 1 = d.key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n92 -POSTHOOK: Input: default@t2_n57 -POSTHOOK: Input: default@t3_n21 -POSTHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: (key is not null and val is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key is not null and val is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - TableScan - alias: b - filterExpr: (key is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key is not null and (key + 1) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col5 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: val is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: val (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: val (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: string) - 1 val (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: (_col0 + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (_col0 + 1) (type: double) - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) - TableScan - alias: d - filterExpr: (key + 1) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (key + 1) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: (key + 1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (key + 1) (type: double) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: key (type: string), val (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 (_col0 + 1) (type: double) - 1 (key + 1) (type: double) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON a.val = c.val - JOIN T4_n10 d ON a.key + 1 = d.key + 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n92 -PREHOOK: Input: default@t2_n57 -PREHOOK: Input: default@t3_n21 -PREHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * -FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key - JOIN T3_n21 c ON a.val = c.val - JOIN T4_n10 d ON a.key + 1 = d.key + 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n92 -POSTHOOK: Input: default@t2_n57 -POSTHOOK: Input: default@t3_n21 -POSTHOOK: Input: default@t4_n10 -#### A masked pattern was here #### -2 12 2 22 2 12 2 12 diff --git ql/src/test/results/clientpositive/join_reorder4.q.out ql/src/test/results/clientpositive/join_reorder4.q.out deleted file mode 100644 index f57e897443..0000000000 --- ql/src/test/results/clientpositive/join_reorder4.q.out +++ /dev/null @@ -1,438 +0,0 @@ -PREHOOK: query: CREATE TABLE T1_n134(key1 STRING, val1 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T1_n134 -POSTHOOK: query: CREATE TABLE T1_n134(key1 STRING, val1 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T1_n134 -PREHOOK: query: CREATE TABLE T2_n80(key2 STRING, val2 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T2_n80 -POSTHOOK: query: CREATE TABLE T2_n80(key2 STRING, val2 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T2_n80 -PREHOOK: query: CREATE TABLE T3_n32(key3 STRING, val3 STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@T3_n32 -POSTHOOK: query: CREATE TABLE T3_n32(key3 STRING, val3 STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@T3_n32 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n134 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t1_n134 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n134 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t1_n134 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n80 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t2_n80 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n80 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t2_n80 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n32 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@t3_n32 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n32 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@t3_n32 -PREHOOK: query: explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val1 (type: string) - TableScan - alias: b - filterExpr: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key2 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val2 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key1 (type: string) - 1 key2 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key3 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key3 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val3 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 key3 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -2 12 2 22 2 12 -PREHOOK: query: explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val1 (type: string) - TableScan - alias: b - filterExpr: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key2 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val2 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key1 (type: string) - 1 key2 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key3 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key3 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val3 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 key3 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -2 12 2 22 2 12 -PREHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key1 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val1 (type: string) - TableScan - alias: b - filterExpr: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key2 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val2 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key1 (type: string) - 1 key2 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) - TableScan - alias: c - filterExpr: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key3 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key3 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: val3 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 key3 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -PREHOOK: type: QUERY -PREHOOK: Input: default@t1_n134 -PREHOOK: Input: default@t2_n80 -PREHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -POSTHOOK: query: select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@t1_n134 -POSTHOOK: Input: default@t2_n80 -POSTHOOK: Input: default@t3_n32 -#### A masked pattern was here #### -2 12 2 22 2 12 diff --git ql/src/test/results/clientpositive/join_star.q.out ql/src/test/results/clientpositive/join_star.q.out deleted file mode 100644 index 5702ff0be4..0000000000 --- ql/src/test/results/clientpositive/join_star.q.out +++ /dev/null @@ -1,946 +0,0 @@ -PREHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@fact -POSTHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@fact -PREHOOK: query: create table dim1(f1 int, f2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim1 -POSTHOOK: query: create table dim1(f1 int, f2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim1 -PREHOOK: query: create table dim2(f3 int, f4 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim2 -POSTHOOK: query: create table dim2(f3 int, f4 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim2 -PREHOOK: query: create table dim3(f5 int, f6 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim3 -POSTHOOK: query: create table dim3(f5 int, f6 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim3 -PREHOOK: query: create table dim4(f7 int, f8 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim4 -POSTHOOK: query: create table dim4(f7 int, f8 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim4 -PREHOOK: query: create table dim5(f9 int, f10 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim5 -POSTHOOK: query: create table dim5(f9 int, f10 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim5 -PREHOOK: query: create table dim6(f11 int, f12 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim6 -POSTHOOK: query: create table dim6(f11 int, f12 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim6 -PREHOOK: query: create table dim7(f13 int, f14 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dim7 -POSTHOOK: query: create table dim7(f13 int, f14 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dim7 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@fact -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@fact -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim1 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim1 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim2 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim2 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim3 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim3 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim4 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim4 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim5 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim5 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim6 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim6 -PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7 -PREHOOK: type: LOAD -#### A masked pattern was here #### -PREHOOK: Output: default@dim7 -POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7 -POSTHOOK: type: LOAD -#### A masked pattern was here #### -POSTHOOK: Output: default@dim7 -PREHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:dim1 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:dim1 - TableScan - alias: dim1 - filterExpr: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f1 (type: int), f2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: fact - filterExpr: d1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: d1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: m1 (type: int), m2 (type: int), d1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -1 2 1 -11 12 1 -21 22 1 -31 32 1 -41 42 1 -51 52 3 -61 62 3 -71 72 3 -81 82 3 -91 92 3 -PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-5 depends on stages: Stage-7 - Stage-0 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-7 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:dim1 - Fetch Operator - limit: -1 - $hdt$_2:dim2 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:dim1 - TableScan - alias: dim1 - filterExpr: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f1 (type: int), f2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - $hdt$_2:dim2 - TableScan - alias: dim2 - filterExpr: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f3 (type: int), f4 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col3 (type: int) - 1 _col0 (type: int) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: fact - filterExpr: (d1 is not null and d2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (d1 is not null and d2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: m1 (type: int), m2 (type: int), d1 (type: int), d2 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3, _col5 - Statistics: Num rows: 1 Data size: 17 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col3 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7 - Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col7 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -1 2 1 2 -11 12 1 2 -21 22 1 2 -31 32 1 2 -41 42 1 2 -51 52 3 4 -61 62 3 4 -71 72 3 4 -81 82 3 4 -91 92 3 4 -PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-5 depends on stages: Stage-7 - Stage-0 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-7 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:dim1 - Fetch Operator - limit: -1 - $hdt$_2:fact - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:dim1 - TableScan - alias: dim1 - filterExpr: (f1 is not null and f2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (f1 is not null and f2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f1 (type: int), f2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col1 (type: int) - 1 _col0 (type: int) - $hdt$_2:fact - TableScan - alias: fact - filterExpr: d1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: d1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: m1 (type: int), m2 (type: int), d1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col0 (type: int) - 1 _col2 (type: int) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: dim2 - filterExpr: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f3 (type: int), f4 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col2 (type: int) - outputColumnNames: _col1, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col4 (type: int), _col5 (type: int), _col1 (type: int), _col3 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -1 2 1 1 -11 12 1 1 -21 22 1 1 -31 32 1 1 -41 42 1 1 -51 52 3 3 -61 62 3 3 -71 72 3 3 -81 82 3 3 -91 92 3 3 -PREHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-7 is a root stage - Stage-5 depends on stages: Stage-7 - Stage-0 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-7 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:dim1 - Fetch Operator - limit: -1 - $hdt$_2:dim2 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:dim1 - TableScan - alias: dim1 - filterExpr: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f1 (type: int), f2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - $hdt$_2:dim2 - TableScan - alias: dim2 - filterExpr: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f3 (type: int), f4 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col4 (type: int) - 1 _col0 (type: int) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: fact - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: m1 (type: int), m2 (type: int), d1 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col4 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col4, _col6 - Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col6 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -1 2 1 1 -11 12 1 1 -21 22 1 1 -31 32 1 1 -41 42 1 1 -51 52 3 3 -61 62 3 3 -71 72 3 3 -81 82 3 3 -91 92 3 3 -PREHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14 - from fact - Left outer join dim1 on fact.d1= dim1.f1 - Left outer join dim2 on dim1.f2 = dim2.f3 - Left outer Join dim3 on fact.d2= dim3.f5 - Left outer Join dim4 on dim3.f6= dim4.f7 - Left outer join dim5 on dim4.f8= dim5.f9 - Left outer Join dim6 on dim3.f6= dim6.f11 - Left outer Join dim7 on dim6.f12 = dim7.f13 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@dim3 -PREHOOK: Input: default@dim4 -PREHOOK: Input: default@dim5 -PREHOOK: Input: default@dim6 -PREHOOK: Input: default@dim7 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14 - from fact - Left outer join dim1 on fact.d1= dim1.f1 - Left outer join dim2 on dim1.f2 = dim2.f3 - Left outer Join dim3 on fact.d2= dim3.f5 - Left outer Join dim4 on dim3.f6= dim4.f7 - Left outer join dim5 on dim4.f8= dim5.f9 - Left outer Join dim6 on dim3.f6= dim6.f11 - Left outer Join dim7 on dim6.f12 = dim7.f13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@dim3 -POSTHOOK: Input: default@dim4 -POSTHOOK: Input: default@dim5 -POSTHOOK: Input: default@dim6 -POSTHOOK: Input: default@dim7 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-22 is a root stage - Stage-15 depends on stages: Stage-22 - Stage-0 depends on stages: Stage-15 - -STAGE PLANS: - Stage: Stage-22 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:dim1 - Fetch Operator - limit: -1 - $hdt$_2:dim2 - Fetch Operator - limit: -1 - $hdt$_3:dim3 - Fetch Operator - limit: -1 - $hdt$_4:dim4 - Fetch Operator - limit: -1 - $hdt$_5:dim5 - Fetch Operator - limit: -1 - $hdt$_6:dim6 - Fetch Operator - limit: -1 - $hdt$_7:dim7 - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:dim1 - TableScan - alias: dim1 - filterExpr: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f1 (type: int), f2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - $hdt$_2:dim2 - TableScan - alias: dim2 - filterExpr: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f3 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f3 (type: int), f4 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col5 (type: int) - 1 _col0 (type: int) - $hdt$_3:dim3 - TableScan - alias: dim3 - filterExpr: f5 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f5 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f5 (type: int), f6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col3 (type: int) - 1 _col0 (type: int) - $hdt$_4:dim4 - TableScan - alias: dim4 - filterExpr: f7 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f7 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f7 (type: int), f8 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col9 (type: int) - 1 _col0 (type: int) - $hdt$_5:dim5 - TableScan - alias: dim5 - filterExpr: f9 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f9 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f9 (type: int), f10 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col11 (type: int) - 1 _col0 (type: int) - $hdt$_6:dim6 - TableScan - alias: dim6 - filterExpr: f11 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f11 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f11 (type: int), f12 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col9 (type: int) - 1 _col0 (type: int) - $hdt$_7:dim7 - TableScan - alias: dim7 - filterExpr: f13 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: f13 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: f13 (type: int), f14 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - HashTable Sink Operator - keys: - 0 _col15 (type: int) - 1 _col0 (type: int) - - Stage: Stage-15 - Map Reduce - Map Operator Tree: - TableScan - alias: fact - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: m1 (type: int), m2 (type: int), d1 (type: int), d2 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col2 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3, _col5 - Statistics: Num rows: 1 Data size: 17 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col5 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3, _col5, _col7 - Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col3 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7, _col9 - Statistics: Num rows: 1 Data size: 19 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col9 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11 - Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col11 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13 - Statistics: Num rows: 1 Data size: 22 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col9 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13, _col15 - Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE Column stats: NONE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col15 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13, _col15, _col17 - Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col7 (type: int), _col9 (type: int), _col11 (type: int), _col13 (type: int), _col15 (type: int), _col17 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14 - from fact - Left outer join dim1 on fact.d1= dim1.f1 - Left outer join dim2 on dim1.f2 = dim2.f3 - Left outer Join dim3 on fact.d2= dim3.f5 - Left outer Join dim4 on dim3.f6= dim4.f7 - Left outer join dim5 on dim4.f8= dim5.f9 - Left outer Join dim6 on dim3.f6= dim6.f11 - Left outer Join dim7 on dim6.f12 = dim7.f13 -PREHOOK: type: QUERY -PREHOOK: Input: default@dim1 -PREHOOK: Input: default@dim2 -PREHOOK: Input: default@dim3 -PREHOOK: Input: default@dim4 -PREHOOK: Input: default@dim5 -PREHOOK: Input: default@dim6 -PREHOOK: Input: default@dim7 -PREHOOK: Input: default@fact -#### A masked pattern was here #### -POSTHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14 - from fact - Left outer join dim1 on fact.d1= dim1.f1 - Left outer join dim2 on dim1.f2 = dim2.f3 - Left outer Join dim3 on fact.d2= dim3.f5 - Left outer Join dim4 on dim3.f6= dim4.f7 - Left outer join dim5 on dim4.f8= dim5.f9 - Left outer Join dim6 on dim3.f6= dim6.f11 - Left outer Join dim7 on dim6.f12 = dim7.f13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@dim1 -POSTHOOK: Input: default@dim2 -POSTHOOK: Input: default@dim3 -POSTHOOK: Input: default@dim4 -POSTHOOK: Input: default@dim5 -POSTHOOK: Input: default@dim6 -POSTHOOK: Input: default@dim7 -POSTHOOK: Input: default@fact -#### A masked pattern was here #### -1 2 1 1 2 2 2 2 2 -11 12 1 1 2 2 2 2 2 -21 22 1 1 2 2 2 2 2 -31 32 1 1 2 2 2 2 2 -41 42 1 1 2 2 2 2 2 -51 52 3 3 4 4 4 4 4 -61 62 3 3 4 4 4 4 4 -71 72 3 3 4 4 4 4 4 -81 82 3 3 4 4 4 4 4 -91 92 3 3 4 4 4 4 4 diff --git ql/src/test/results/clientpositive/join_thrift.q.out ql/src/test/results/clientpositive/join_thrift.q.out deleted file mode 100644 index 2ae592367b..0000000000 --- ql/src/test/results/clientpositive/join_thrift.q.out +++ /dev/null @@ -1,119 +0,0 @@ -PREHOOK: query: DESCRIBE src_thrift -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@src_thrift -POSTHOOK: query: DESCRIBE src_thrift -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@src_thrift -aint int from deserializer -astring string from deserializer -lint array from deserializer -lstring array from deserializer -lintstring array> from deserializer -mstringstring map from deserializer -attributes map,map>>>> from deserializer -unionfield1 uniontype,map> from deserializer -unionfield2 uniontype,map> from deserializer -unionfield3 uniontype,map> from deserializer -PREHOOK: query: EXPLAIN -SELECT s1.aint, s2.lintstring -FROM src_thrift s1 -JOIN src_thrift s2 -ON s1.aint = s2.aint -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT s1.aint, s2.lintstring -FROM src_thrift s1 -JOIN src_thrift s2 -ON s1.aint = s2.aint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: s1 - filterExpr: aint is not null (type: boolean) - Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: aint is not null (type: boolean) - Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: aint (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: aint (type: int) - Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE - TableScan - alias: s2 - filterExpr: aint is not null (type: boolean) - Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: aint is not null (type: boolean) - Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: aint (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: aint (type: int) - Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE - value expressions: lintstring (type: array>) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 aint (type: int) - 1 aint (type: int) - outputColumnNames: _col0, _col17 - Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int), _col17 (type: array>) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT s1.aint, s2.lintstring -FROM src_thrift s1 -JOIN src_thrift s2 -ON s1.aint = s2.aint -PREHOOK: type: QUERY -PREHOOK: Input: default@src_thrift -#### A masked pattern was here #### -POSTHOOK: query: SELECT s1.aint, s2.lintstring -FROM src_thrift s1 -JOIN src_thrift s2 -ON s1.aint = s2.aint -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src_thrift -#### A masked pattern was here #### --1461153973 [{"myint":49,"mystring":"343","underscore_int":7}] --1952710710 [{"myint":25,"mystring":"125","underscore_int":5}] --734328909 [{"myint":16,"mystring":"64","underscore_int":4}] --751827638 [{"myint":4,"mystring":"8","underscore_int":2}] -0 NULL -1244525190 [{"myint":36,"mystring":"216","underscore_int":6}] -1638581578 [{"myint":64,"mystring":"512","underscore_int":8}] -1712634731 [{"myint":0,"mystring":"0","underscore_int":0}] -336964413 [{"myint":81,"mystring":"729","underscore_int":9}] -465985200 [{"myint":1,"mystring":"1","underscore_int":1}] -477111222 [{"myint":9,"mystring":"27","underscore_int":3}] diff --git ql/src/test/results/clientpositive/join_view.q.out ql/src/test/results/clientpositive/join_view.q.out deleted file mode 100644 index eb8e7717dd..0000000000 --- ql/src/test/results/clientpositive/join_view.q.out +++ /dev/null @@ -1,147 +0,0 @@ -PREHOOK: query: drop table invites -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table invites -POSTHOOK: type: DROPTABLE -PREHOOK: query: drop table invites2 -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table invites2 -POSTHOOK: type: DROPTABLE -PREHOOK: query: create table invites (foo int, bar string) partitioned by (ds string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@invites -POSTHOOK: query: create table invites (foo int, bar string) partitioned by (ds string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@invites -PREHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@invites2 -POSTHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@invites2 -PREHOOK: query: create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@invites -PREHOOK: Input: default@invites2 -PREHOOK: Output: database:default -PREHOOK: Output: default@v -POSTHOOK: query: create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@invites -POSTHOOK: Input: default@invites2 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v -POSTHOOK: Lineage: v.bar SIMPLE [(invites)invites.FieldSchema(name:bar, type:string, comment:null), ] -POSTHOOK: Lineage: v.ds SIMPLE [(invites2)invites2.FieldSchema(name:ds, type:string, comment:null), ] -POSTHOOK: Lineage: v.foo SIMPLE [(invites2)invites2.FieldSchema(name:foo, type:int, comment:null), ] -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain select * from v where ds='2011-09-01' -PREHOOK: type: QUERY -PREHOOK: Input: default@invites -PREHOOK: Input: default@invites2 -PREHOOK: Input: default@v -#### A masked pattern was here #### -POSTHOOK: query: explain select * from v where ds='2011-09-01' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@invites -POSTHOOK: Input: default@invites2 -POSTHOOK: Input: default@v -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: invites - filterExpr: (ds = '2011-09-01') (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (ds = '2011-09-01') (type: boolean) - Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: bar (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: invites2 - filterExpr: (ds = '2011-09-01') (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (ds = '2011-09-01') (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: foo (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: int), '2011-09-01' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: drop view v -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@v -PREHOOK: Output: default@v -POSTHOOK: query: drop view v -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@v -POSTHOOK: Output: default@v -PREHOOK: query: drop table invites -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@invites -PREHOOK: Output: default@invites -POSTHOOK: query: drop table invites -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@invites -POSTHOOK: Output: default@invites -PREHOOK: query: drop table invites2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@invites2 -PREHOOK: Output: default@invites2 -POSTHOOK: query: drop table invites2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@invites2 -POSTHOOK: Output: default@invites2 diff --git ql/src/test/results/clientpositive/lateral_view_cp.q.out ql/src/test/results/clientpositive/lateral_view_cp.q.out deleted file mode 100644 index 233b3bcfc1..0000000000 --- ql/src/test/results/clientpositive/lateral_view_cp.q.out +++ /dev/null @@ -1,164 +0,0 @@ -PREHOOK: query: drop table array_valued_src -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table array_valued_src -POSTHOOK: type: DROPTABLE -PREHOOK: query: create table array_valued_src (key string, value array) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@array_valued_src -POSTHOOK: query: create table array_valued_src (key string, value array) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@array_valued_src -PREHOOK: query: insert overwrite table array_valued_src select key, array(value) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@array_valued_src -POSTHOOK: query: insert overwrite table array_valued_src select key, array(value) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@array_valued_src -POSTHOOK: Lineage: array_valued_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: array_valued_src.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val -PREHOOK: type: QUERY -PREHOOK: Input: default@array_valued_src -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@array_valued_src -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 475 Data size: 951428 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 475 Data size: 951428 Basic stats: COMPLETE Column stats: NONE - value expressions: value (type: array) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col6 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col6 (type: array) - outputColumnNames: _col1 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Lateral View Forward - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Select Operator - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Lateral View Join Operator - outputColumnNames: _col2 - Statistics: Num rows: 1100 Data size: 95700 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(_col2) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Select Operator - expressions: _col1 (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - UDTF Operator - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - function name: explode - Lateral View Join Operator - outputColumnNames: _col2 - Statistics: Num rows: 1100 Data size: 95700 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(_col2) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val -PREHOOK: type: QUERY -PREHOOK: Input: default@array_valued_src -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val -POSTHOOK: type: QUERY -POSTHOOK: Input: default@array_valued_src -POSTHOOK: Input: default@src -#### A masked pattern was here #### -1028 diff --git ql/src/test/results/clientpositive/lateral_view_explode2.q.out ql/src/test/results/clientpositive/lateral_view_explode2.q.out deleted file mode 100644 index 324c271eac..0000000000 --- ql/src/test/results/clientpositive/lateral_view_explode2.q.out +++ /dev/null @@ -1,108 +0,0 @@ -PREHOOK: query: CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2' -PREHOOK: type: CREATEFUNCTION -PREHOOK: Output: explode2 -POSTHOOK: query: CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2' -POSTHOOK: type: CREATEFUNCTION -POSTHOOK: Output: explode2 -PREHOOK: query: EXPLAIN SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col5 (type: int), _col6 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col5 (type: int), _col6 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 3 - Processor Tree: - ListSink - -PREHOOK: query: SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -1 1 -2 2 -3 3 -PREHOOK: query: DROP TEMPORARY FUNCTION explode2 -PREHOOK: type: DROPFUNCTION -PREHOOK: Output: explode2 -POSTHOOK: query: DROP TEMPORARY FUNCTION explode2 -POSTHOOK: type: DROPFUNCTION -POSTHOOK: Output: explode2 diff --git ql/src/test/results/clientpositive/lateral_view_noalias.q.out ql/src/test/results/clientpositive/lateral_view_noalias.q.out deleted file mode 100644 index e820ab9747..0000000000 --- ql/src/test/results/clientpositive/lateral_view_noalias.q.out +++ /dev/null @@ -1,325 +0,0 @@ -PREHOOK: query: EXPLAIN SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-0 - Fetch Operator - limit: 2 - Processor Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - ListSink - Select Operator - expressions: map('key1':100,'key2':200) (type: map) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - ListSink - -PREHOOK: query: SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -key1 100 -key2 200 -PREHOOK: query: EXPLAIN SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-0 is a root stage - -STAGE PLANS: - Stage: Stage-0 - Fetch Operator - limit: 2 - Processor Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: map('key1':100,'key2':200) (type: map) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Select Operator - expressions: key (type: string), value (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - ListSink - -PREHOOK: query: SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -key1 100 -key2 200 -PREHOOK: query: create view lv_noalias as SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_noalias -POSTHOOK: query: create view lv_noalias as SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_noalias -POSTHOOK: Lineage: lv_noalias.key SCRIPT [] -POSTHOOK: Lineage: lv_noalias.value SCRIPT [] -PREHOOK: query: explain select * from lv_noalias a join lv_noalias b on a.key=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_noalias -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from lv_noalias a join lv_noalias b on a.key=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_noalias -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: int) - Select Operator - expressions: map('key1':100,'key2':200) (type: map) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: int) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: int) - Select Operator - expressions: map('key1':100,'key2':200) (type: map) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5, _col6 - Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: string), _col6 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: int) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col0 is not null (type: boolean) - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from lv_noalias a join lv_noalias b on a.key=b.key -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_noalias -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from lv_noalias a join lv_noalias b on a.key=b.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_noalias -POSTHOOK: Input: default@src -#### A masked pattern was here #### -key1 100 key1 100 -key2 200 key2 200 diff --git ql/src/test/results/clientpositive/lateral_view_onview.q.out ql/src/test/results/clientpositive/lateral_view_onview.q.out deleted file mode 100644 index 2157e91e8d..0000000000 --- ql/src/test/results/clientpositive/lateral_view_onview.q.out +++ /dev/null @@ -1,856 +0,0 @@ -PREHOOK: query: CREATE TABLE lv_table_n0( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_table_n0 -POSTHOOK: query: CREATE TABLE lv_table_n0( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_table_n0 -PREHOOK: query: INSERT OVERWRITE TABLE lv_table_n0 SELECT 'abc ', array(1,2,3), 100, 't' FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@lv_table_n0 -POSTHOOK: query: INSERT OVERWRITE TABLE lv_table_n0 SELECT 'abc ', array(1,2,3), 100, 't' FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@lv_table_n0 -POSTHOOK: Lineage: lv_table_n0.c1 SIMPLE [] -POSTHOOK: Lineage: lv_table_n0.c2 EXPRESSION [] -POSTHOOK: Lineage: lv_table_n0.c3 SIMPLE [] -POSTHOOK: Lineage: lv_table_n0.c4 EXPRESSION [] -PREHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n0 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_view -POSTHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n0 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_view -POSTHOOK: Lineage: lv_view.c1 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c1, type:string, comment:null), ] -POSTHOOK: Lineage: lv_view.c2 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c2, type:array, comment:null), ] -POSTHOOK: Lineage: lv_view.c3 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c3, type:int, comment:null), ] -POSTHOOK: Lineage: lv_view.c4 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c4, type:char(1), comment:null), ] -PREHOOK: query: EXPLAIN SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n0 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: c1 (type: string), c2 (type: array), c3 (type: int), c4 (type: char(1)) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Lateral View Forward - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col4 (type: int) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - UDTF Operator - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - function name: explode - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col4 (type: int) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: array), VALUE._col1 (type: int), VALUE._col2 (type: char(1)), KEY.reducesinkkey1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col4 (type: int) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: array), VALUE._col1 (type: int), VALUE._col2 (type: char(1)), KEY.reducesinkkey1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 1 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n0 - Statistics: Num rows: 500 Data size: 8500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 3 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n0 - Statistics: Num rows: 500 Data size: 8500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array('a','b','c') (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array('a','b','c') (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 9 - Processor Tree: - ListSink - -PREHOOK: query: EXPLAIN SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n0 - Statistics: Num rows: 500 Data size: 8500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1000 Data size: 48000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: _col4 (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(array(1,2,3)) (type: array>) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 24000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 24000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 1000 Data size: 48000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col5 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: _col4 (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col5 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 3 - Processor Tree: - ListSink - -PREHOOK: query: SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -abc [1,2,3] 100 t 1 -PREHOOK: query: SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -1 -2 -3 -PREHOOK: query: SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -1 a -1 b -1 c -2 a -2 b -2 c -3 a -3 b -3 c -PREHOOK: query: SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -1 -2 -3 -PREHOOK: query: EXPLAIN -SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN -SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n0 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: c1 (type: string), c2 (type: array), c3 (type: int), c4 (type: char(1)) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Lateral View Forward - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: size(_col1) (type: int), _col2 (type: int), trim(_col0) (type: string), _col3 (type: char(1)), _col4 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 1122 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1122 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - UDTF Operator - Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE - function name: explode - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: size(_col1) (type: int), _col2 (type: int), trim(_col0) (type: string), _col3 (type: char(1)), _col4 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 3 - Statistics: Num rows: 3 Data size: 1122 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1122 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 3 - Processor Tree: - ListSink - -PREHOOK: query: SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n0 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n0 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -3 100 abc t 1 -3 100 abc t 2 -3 100 abc t 3 -PREHOOK: query: CREATE TABLE lv_table1_n0( c1 STRING, c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_table1_n0 -POSTHOOK: query: CREATE TABLE lv_table1_n0( c1 STRING, c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_table1_n0 -PREHOOK: query: CREATE TABLE lv_table2( c1 STRING, c2 ARRAY) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_table2 -POSTHOOK: query: CREATE TABLE lv_table2( c1 STRING, c2 ARRAY) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_table2 -PREHOOK: query: INSERT OVERWRITE TABLE lv_table1_n0 SELECT 'abc ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@lv_table1_n0 -POSTHOOK: query: INSERT OVERWRITE TABLE lv_table1_n0 SELECT 'abc ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@lv_table1_n0 -POSTHOOK: Lineage: lv_table1_n0.c1 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c10 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c11 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c12 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c13 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c3 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c4 EXPRESSION [] -POSTHOOK: Lineage: lv_table1_n0.c5 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c6 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c7 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c8 SIMPLE [] -POSTHOOK: Lineage: lv_table1_n0.c9 SIMPLE [] -PREHOOK: query: INSERT OVERWRITE TABLE lv_table2 SELECT 'abc ', array(1,2,3) FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@lv_table2 -POSTHOOK: query: INSERT OVERWRITE TABLE lv_table2 SELECT 'abc ', array(1,2,3) FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@lv_table2 -POSTHOOK: Lineage: lv_table2.c1 SIMPLE [] -POSTHOOK: Lineage: lv_table2.c2 EXPRESSION [] -PREHOOK: query: EXPLAIN WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table1_n0 -PREHOOK: Input: default@lv_table2 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table1_n0 -POSTHOOK: Input: default@lv_table2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table1_n0 - filterExpr: c1 is not null (type: boolean) - Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: c1 is not null (type: boolean) - Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: c1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: c1 (type: string) - Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: c3 (type: int), c4 (type: char(1)), c5 (type: string), c6 (type: string), c7 (type: string), c8 (type: string), c9 (type: string), c10 (type: string), c11 (type: string), c12 (type: string), c13 (type: string) - TableScan - alias: lv_table2 - filterExpr: c1 is not null (type: boolean) - Statistics: Num rows: 500 Data size: 144704 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: c1 is not null (type: boolean) - Statistics: Num rows: 475 Data size: 137468 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: c1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: c1 (type: string) - Statistics: Num rows: 475 Data size: 137468 Basic stats: COMPLETE Column stats: NONE - value expressions: c2 (type: array) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 c1 (type: string) - 1 c1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col16 - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col16 (type: array) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - Lateral View Forward - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Select Operator - expressions: _col12 (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - UDTF Operator - Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE - function name: explode - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col13 (type: int) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: int), VALUE._col1 (type: char(1)), VALUE._col2 (type: string), VALUE._col3 (type: string), VALUE._col4 (type: string), VALUE._col5 (type: string), VALUE._col6 (type: string), VALUE._col7 (type: string), VALUE._col8 (type: string), VALUE._col9 (type: string), VALUE._col10 (type: string), VALUE._col11 (type: array), KEY.reducesinkkey1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col13 (type: int) - null sort order: zz - sort order: ++ - Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE - TopN Hash Memory Usage: 0.1 - value expressions: _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: int), VALUE._col1 (type: char(1)), VALUE._col2 (type: string), VALUE._col3 (type: string), VALUE._col4 (type: string), VALUE._col5 (type: string), VALUE._col6 (type: string), VALUE._col7 (type: string), VALUE._col8 (type: string), VALUE._col9 (type: string), VALUE._col10 (type: string), VALUE._col11 (type: array), KEY.reducesinkkey1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 - Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 1 - Processor Tree: - ListSink - -PREHOOK: query: WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table1_n0 -PREHOOK: Input: default@lv_table2 -#### A masked pattern was here #### -POSTHOOK: query: WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table1_n0 -POSTHOOK: Input: default@lv_table2 -#### A masked pattern was here #### -abc 100 t test test test test test test test test test [1,2,3] 1 diff --git ql/src/test/results/clientpositive/lateral_view_onview2.q.out ql/src/test/results/clientpositive/lateral_view_onview2.q.out deleted file mode 100644 index 55f869bc11..0000000000 --- ql/src/test/results/clientpositive/lateral_view_onview2.q.out +++ /dev/null @@ -1,175 +0,0 @@ -PREHOOK: query: CREATE TABLE lv_table_n1( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_table_n1 -POSTHOOK: query: CREATE TABLE lv_table_n1( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_table_n1 -PREHOOK: query: INSERT OVERWRITE TABLE lv_table_n1 SELECT 'abc ', array(1,2,3), 100, 't' FROM src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@lv_table_n1 -POSTHOOK: query: INSERT OVERWRITE TABLE lv_table_n1 SELECT 'abc ', array(1,2,3), 100, 't' FROM src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@lv_table_n1 -POSTHOOK: Lineage: lv_table_n1.c1 SIMPLE [] -POSTHOOK: Lineage: lv_table_n1.c2 EXPRESSION [] -POSTHOOK: Lineage: lv_table_n1.c3 SIMPLE [] -POSTHOOK: Lineage: lv_table_n1.c4 EXPRESSION [] -PREHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n1 -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@lv_table_n1 -PREHOOK: Output: database:default -PREHOOK: Output: default@lv_view -POSTHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n1 -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@lv_table_n1 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@lv_view -POSTHOOK: Lineage: lv_view.c1 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c1, type:string, comment:null), ] -POSTHOOK: Lineage: lv_view.c2 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c2, type:array, comment:null), ] -POSTHOOK: Lineage: lv_view.c3 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c3, type:int, comment:null), ] -POSTHOOK: Lineage: lv_view.c4 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c4, type:char(1), comment:null), ] -PREHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -PREHOOK: type: QUERY -PREHOOK: Input: default@lv_table_n1 -PREHOOK: Input: default@lv_view -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@lv_table_n1 -POSTHOOK: Input: default@lv_view -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: lv_table_n1 - Statistics: Num rows: 500 Data size: 8500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array('a','b','c') (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 1000 Data size: 32000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int) - outputColumnNames: _col4 - Statistics: Num rows: 1000 Data size: 56000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array('a','b','c') (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 1000 Data size: 592000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col4, _col5 - Statistics: Num rows: 2000 Data size: 648000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: int), _col5 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 16000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 9 - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 9 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 9 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/lateral_view_outer.q.out ql/src/test/results/clientpositive/lateral_view_outer.q.out deleted file mode 100644 index 8fe15883fa..0000000000 --- ql/src/test/results/clientpositive/lateral_view_outer.q.out +++ /dev/null @@ -1,294 +0,0 @@ -PREHOOK: query: explain -select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 261000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 281000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 281000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array() (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 20000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 20000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - outer lateral view: true - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 281000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 281000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -238 val_238 NULL -86 val_86 NULL -311 val_311 NULL -27 val_27 NULL -165 val_165 NULL -409 val_409 NULL -255 val_255 NULL -278 val_278 NULL -98 val_98 NULL -484 val_484 NULL -PREHOOK: query: explain -select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 261000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 285000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 285000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(4,5) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 24000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 24000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - outer lateral view: true - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 285000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 285000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -238 val_238 4 -238 val_238 5 -86 val_86 4 -86 val_86 5 -311 val_311 4 -311 val_311 5 -27 val_27 4 -27 val_27 5 -165 val_165 4 -165 val_165 5 -PREHOOK: query: create table array_valued as select key, if (key > 300, array(value, value), null) as value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@array_valued -POSTHOOK: query: create table array_valued as select key, if (key > 300, array(value, value), null) as value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@array_valued -POSTHOOK: Lineage: array_valued.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: array_valued.value EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@array_valued -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@array_valued -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: array_valued - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - Lateral View Forward - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), value (type: array) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 2003008 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: array), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 2003008 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 20030 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 20030 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: value (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - UDTF Operator - Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE - function name: explode - outer lateral view: true - Lateral View Join Operator - outputColumnNames: _col0, _col1, _col5 - Statistics: Num rows: 1000 Data size: 2003008 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: array), _col5 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 2003008 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 20030 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 20030 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@array_valued -#### A masked pattern was here #### -POSTHOOK: query: select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@array_valued -#### A masked pattern was here #### -238 NULL NULL -86 NULL NULL -311 ["val_311","val_311"] val_311 -311 ["val_311","val_311"] val_311 -27 NULL NULL -165 NULL NULL -409 ["val_409","val_409"] val_409 -409 ["val_409","val_409"] val_409 -255 NULL NULL -278 NULL NULL diff --git ql/src/test/results/clientpositive/lateral_view_ppd.q.out ql/src/test/results/clientpositive/lateral_view_ppd.q.out deleted file mode 100644 index 60d1629c01..0000000000 --- ql/src/test/results/clientpositive/lateral_view_ppd.q.out +++ /dev/null @@ -1,525 +0,0 @@ -PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '0') (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '0') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: value - Statistics: Num rows: 2 Data size: 1040 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -val_0 1 -val_0 2 -val_0 3 -val_0 1 -val_0 2 -val_0 3 -val_0 1 -val_0 2 -val_0 3 -PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '0') (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '0') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: value - Statistics: Num rows: 2 Data size: 1040 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 3 Data size: 1096 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), 1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Filter Operator - predicate: (col = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 3 Data size: 1096 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), 1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -val_0 1 -val_0 1 -val_0 1 -PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: ((ds = '2008-04-08') and (hr = '12')) (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: value - Statistics: Num rows: 500 Data size: 307500 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col7 - Statistics: Num rows: 1000 Data size: 335500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col7 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 12 - Statistics: Num rows: 12 Data size: 1092 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 1092 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col1, _col7 - Statistics: Num rows: 1000 Data size: 335500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col7 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 12 - Statistics: Num rows: 12 Data size: 1092 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 12 Data size: 1092 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 12 - Processor Tree: - ListSink - -PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -val_238 1 -val_238 2 -val_238 3 -val_86 1 -val_86 2 -val_86 3 -val_311 1 -val_311 2 -val_311 3 -val_27 1 -val_27 2 -val_27 3 -PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '0') (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '0') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: value - Statistics: Num rows: 2 Data size: 1040 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1776 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5, _col6 - Statistics: Num rows: 8 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 224 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 4 Data size: 224 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col1, _col5, _col6 - Statistics: Num rows: 8 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 4 Data size: 1152 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col1, _col5 - Statistics: Num rows: 4 Data size: 1776 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5, _col6 - Statistics: Num rows: 8 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 4 Data size: 224 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 4 Data size: 224 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col1, _col5, _col6 - Statistics: Num rows: 8 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 8 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -val_0 1 -val_0 1 -val_0 1 -val_0 2 -val_0 2 -val_0 2 -val_0 3 -val_0 3 -val_0 3 -val_0 1 -val_0 1 -val_0 1 -val_0 2 -val_0 2 -val_0 2 -val_0 3 -val_0 3 -val_0 3 -val_0 1 -val_0 1 -val_0 1 -val_0 2 -val_0 2 -val_0 2 -val_0 3 -val_0 3 -val_0 3 -PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '0') (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '0') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: value - Statistics: Num rows: 2 Data size: 1040 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 3 Data size: 1096 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(1,2,3) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 2 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Filter Operator - predicate: (col > 1) (type: boolean) - Statistics: Num rows: 1 Data size: 56 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col1, _col5 - Statistics: Num rows: 3 Data size: 1096 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col5 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -val_0 2 -val_0 3 -val_0 2 -val_0 3 -val_0 2 -val_0 3 diff --git ql/src/test/results/clientpositive/limit_pushdown2.q.out ql/src/test/results/clientpositive/limit_pushdown2.q.out deleted file mode 100644 index 43d5284518..0000000000 --- ql/src/test/results/clientpositive/limit_pushdown2.q.out +++ /dev/null @@ -1,1236 +0,0 @@ -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 1.0 -10 val_10 11.0 -100 val_100 101.0 -103 val_103 104.0 -104 val_104 105.0 -105 val_105 106.0 -11 val_11 12.0 -111 val_111 112.0 -113 val_113 114.0 -114 val_114 115.0 -116 val_116 117.0 -118 val_118 119.0 -119 val_119 120.0 -12 val_12 13.0 -120 val_120 121.0 -125 val_125 126.0 -126 val_126 127.0 -128 val_128 129.0 -129 val_129 130.0 -131 val_131 132.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key, value desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key, value desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: +- - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key, value desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key, value desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 1.0 -10 val_10 11.0 -100 val_100 101.0 -103 val_103 104.0 -104 val_104 105.0 -105 val_105 106.0 -11 val_11 12.0 -111 val_111 112.0 -113 val_113 114.0 -114 val_114 115.0 -116 val_116 117.0 -118 val_118 119.0 -119 val_119 120.0 -12 val_12 13.0 -120 val_120 121.0 -125 val_125 126.0 -126 val_126 127.0 -128 val_128 129.0 -129 val_129 130.0 -131 val_131 132.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key desc, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by key, value -order by key desc, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key desc, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by key, value -order by key desc, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -98 val_98 99.0 -97 val_97 98.0 -96 val_96 97.0 -95 val_95 96.0 -92 val_92 93.0 -90 val_90 91.0 -9 val_9 10.0 -87 val_87 88.0 -86 val_86 87.0 -85 val_85 86.0 -84 val_84 85.0 -83 val_83 84.0 -82 val_82 83.0 -80 val_80 81.0 -8 val_8 9.0 -78 val_78 79.0 -77 val_77 78.0 -76 val_76 77.0 -74 val_74 75.0 -72 val_72 73.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col1 (type: string), _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 1.0 -10 val_10 11.0 -100 val_100 101.0 -103 val_103 104.0 -104 val_104 105.0 -105 val_105 106.0 -11 val_11 12.0 -111 val_111 112.0 -113 val_113 114.0 -114 val_114 115.0 -116 val_116 117.0 -118 val_118 119.0 -119 val_119 120.0 -12 val_12 13.0 -120 val_120 121.0 -125 val_125 126.0 -126 val_126 127.0 -128 val_128 129.0 -129 val_129 130.0 -131 val_131 132.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key desc, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key desc, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col1 (type: string), _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key desc, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key desc, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -98 val_98 99.0 -97 val_97 98.0 -96 val_96 97.0 -95 val_95 96.0 -92 val_92 93.0 -90 val_90 91.0 -9 val_9 10.0 -87 val_87 88.0 -86 val_86 87.0 -85 val_85 86.0 -84 val_84 85.0 -83 val_83 84.0 -82 val_82 83.0 -80 val_80 81.0 -8 val_8 9.0 -78 val_78 79.0 -77 val_77 78.0 -76 val_76 77.0 -74 val_74 75.0 -72 val_72 73.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key -order by key desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col1 (type: string), _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double), _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by value, key -order by key desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -98 val_98 99.0 -97 val_97 98.0 -96 val_96 97.0 -95 val_95 96.0 -92 val_92 93.0 -90 val_90 91.0 -9 val_9 10.0 -87 val_87 88.0 -86 val_86 87.0 -85 val_85 86.0 -84 val_84 85.0 -83 val_83 84.0 -82 val_82 83.0 -80 val_80 81.0 -8 val_8 9.0 -78 val_78 79.0 -77 val_77 78.0 -76 val_76 77.0 -74 val_74 75.0 -72 val_72 73.0 -PREHOOK: query: explain -select key, value, count(key + 1) as agg1 from src -group by key, value -order by key, value, agg1 limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, count(key + 1) as agg1 from src -group by key, value -order by key, value, agg1 limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, count(key + 1) as agg1 from src -group by key, value -order by key, value, agg1 limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, count(key + 1) as agg1 from src -group by key, value -order by key, value, agg1 limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 3 -10 val_10 1 -100 val_100 2 -103 val_103 2 -104 val_104 2 -105 val_105 1 -11 val_11 1 -111 val_111 1 -113 val_113 2 -114 val_114 1 -116 val_116 1 -118 val_118 2 -119 val_119 3 -12 val_12 2 -120 val_120 2 -125 val_125 2 -126 val_126 1 -128 val_128 3 -129 val_129 2 -131 val_131 1 -PREHOOK: query: explain -select key, value, count(key + 1) as agg1 from src -group by key, value -order by key desc, value, agg1 limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, count(key + 1) as agg1 from src -group by key, value -order by key desc, value, agg1 limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(_col2) - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: -++ - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: bigint) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, count(key + 1) as agg1 from src -group by key, value -order by key desc, value, agg1 limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, count(key + 1) as agg1 from src -group by key, value -order by key desc, value, agg1 limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -98 val_98 2 -97 val_97 2 -96 val_96 1 -95 val_95 2 -92 val_92 1 -90 val_90 3 -9 val_9 1 -87 val_87 1 -86 val_86 1 -85 val_85 1 -84 val_84 2 -83 val_83 2 -82 val_82 1 -80 val_80 1 -8 val_8 1 -78 val_78 1 -77 val_77 1 -76 val_76 2 -74 val_74 1 -72 val_72 2 -PREHOOK: query: explain -select value, avg(key + 1) myavg from src -group by value -order by myavg, value desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select value, avg(key + 1) myavg from src -group by value -order by myavg, value desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1), count(_col1) - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: double), _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), (_col1 / _col2) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: double), _col0 (type: string) - null sort order: zz - sort order: +- - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select value, avg(key + 1) myavg from src -group by value -order by myavg, value desc limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select value, avg(key + 1) myavg from src -group by value -order by myavg, value desc limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -val_0 1.0 -val_2 3.0 -val_4 5.0 -val_5 6.0 -val_8 9.0 -val_9 10.0 -val_10 11.0 -val_11 12.0 -val_12 13.0 -val_15 16.0 -val_17 18.0 -val_18 19.0 -val_19 20.0 -val_20 21.0 -val_24 25.0 -val_26 27.0 -val_27 28.0 -val_28 29.0 -val_30 31.0 -val_33 34.0 -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key with rollup -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by value, key with rollup -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: double), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3, _col4 - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col1 (type: string), _col0 (type: string), (_col3 / _col4) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: explain -select key, value, avg(key + 1) from src -group by rollup(value, key) -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -select key, value, avg(key + 1) from src -group by rollup(value, key) -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col2), count(_col2) - keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: double), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0), count(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3, _col4 - Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col1 (type: string), _col0 (type: string), (_col3 / _col4) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col2 (type: double) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select key, value, avg(key + 1) from src -group by value, key with rollup -order by key, value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value, avg(key + 1) from src -group by value, key with rollup -order by key, value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 val_0 1.0 -10 val_10 11.0 -100 val_100 101.0 -103 val_103 104.0 -104 val_104 105.0 -105 val_105 106.0 -11 val_11 12.0 -111 val_111 112.0 -113 val_113 114.0 -114 val_114 115.0 -116 val_116 117.0 -118 val_118 119.0 -119 val_119 120.0 -12 val_12 13.0 -120 val_120 121.0 -125 val_125 126.0 -126 val_126 127.0 -128 val_128 129.0 -129 val_129 130.0 -131 val_131 132.0 diff --git ql/src/test/results/clientpositive/limit_pushdown_negative.q.out ql/src/test/results/clientpositive/limit_pushdown_negative.q.out deleted file mode 100644 index 916d59fbe9..0000000000 --- ql/src/test/results/clientpositive/limit_pushdown_negative.q.out +++ /dev/null @@ -1,463 +0,0 @@ -PREHOOK: query: explain select * from src a join src b on a.key=b.key limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from src a join src b on a.key=b.key limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 7120 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 7120 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: explain select value, sum(key) as sum from src group by value having sum > 100 limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select value, sum(key) as sum from src group by value having sum > 100 limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(key) - keys: value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: double) - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col1 > 100.0D) (type: boolean) - Statistics: Num rows: 83 Data size: 8217 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: explain select key, L.* from (select * from src order by key) a lateral view explode(array(value, value)) L as v limit 10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select key, L.* from (select * from src order by key) a lateral view explode(array(value, value)) L as v limit 10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 134500 Basic stats: COMPLETE Column stats: COMPLETE - Lateral View Join Operator - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1000 Data size: 1094500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 2004000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 20040 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 20040 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Select Operator - expressions: array(_col1,_col1) (type: array) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 960000 Basic stats: COMPLETE Column stats: COMPLETE - UDTF Operator - Statistics: Num rows: 500 Data size: 960000 Basic stats: COMPLETE Column stats: COMPLETE - function name: explode - Lateral View Join Operator - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1000 Data size: 1094500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 2004000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 20040 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 20040 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 10 - Processor Tree: - ListSink - -PREHOOK: query: CREATE TABLE dest_2(key STRING, c1 INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_2 -POSTHOOK: query: CREATE TABLE dest_2(key STRING, c1 INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_2 -PREHOOK: query: CREATE TABLE dest_3(key STRING, c1 INT) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_3 -POSTHOOK: query: CREATE TABLE dest_3(key STRING, c1 INT) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_3 -PREHOOK: query: EXPLAIN FROM src -INSERT OVERWRITE TABLE dest_2 SELECT value, sum(key) GROUP BY value -INSERT OVERWRITE TABLE dest_3 SELECT value, sum(key) GROUP BY value limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@dest_2 -PREHOOK: Output: default@dest_3 -POSTHOOK: query: EXPLAIN FROM src -INSERT OVERWRITE TABLE dest_2 SELECT value, sum(key) GROUP BY value -INSERT OVERWRITE TABLE dest_3 SELECT value, sum(key) GROUP BY value limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@dest_2 -POSTHOOK: Output: default@dest_3 -STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: value (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: value (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_2 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, c1 - Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1 - Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_2 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, c1 - Column Types: string, int - Table: default.dest_2 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, c1 - Column Types: string, int - Table: default.dest_3 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.3 - value expressions: _col0 (type: string), _col1 (type: double) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_3 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, c1 - Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_3 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - diff --git ql/src/test/results/clientpositive/lineage1.q.out ql/src/test/results/clientpositive/lineage1.q.out deleted file mode 100644 index 3f86a7fb4b..0000000000 --- ql/src/test/results/clientpositive/lineage1.q.out +++ /dev/null @@ -1,327 +0,0 @@ -PREHOOK: query: CREATE TABLE dest_l1(key INT, value STRING) STORED AS TEXTFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@dest_l1 -POSTHOOK: query: CREATE TABLE dest_l1(key INT, value STRING) STORED AS TEXTFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@dest_l1 -PREHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_l1 -SELECT j.* -FROM (SELECT t1.key, p1.value - FROM src1 t1 - LEFT OUTER JOIN src p1 - ON (t1.key = p1.key) - UNION ALL - SELECT t2.key, p2.value - FROM src1 t2 - LEFT OUTER JOIN src p2 - ON (t2.key = p2.key)) j -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_l1 -POSTHOOK: query: EXPLAIN -INSERT OVERWRITE TABLE dest_l1 -SELECT j.* -FROM (SELECT t1.key, p1.value - FROM src1 t1 - LEFT OUTER JOIN src p1 - ON (t1.key = p1.key) - UNION ALL - SELECT t2.key, p2.value - FROM src1 t2 - LEFT OUTER JOIN src p2 - ON (t2.key = p2.key)) j -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_l1 -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-10 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 - Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-10 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: p1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: t1 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Union - Statistics: Num rows: 128 Data size: 18288 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_l1 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - TableScan - Union - Statistics: Num rows: 128 Data size: 18288 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_l1 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_l1 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: int, string - Table: default.dest_l1 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_l1 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_l1 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - alias: p2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: t2 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - -PREHOOK: query: INSERT OVERWRITE TABLE dest_l1 -SELECT j.* -FROM (SELECT t1.key, p1.value - FROM src1 t1 - LEFT OUTER JOIN src p1 - ON (t1.key = p1.key) - UNION ALL - SELECT t2.key, p2.value - FROM src1 t2 - LEFT OUTER JOIN src p2 - ON (t2.key = p2.key)) j -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@src1 -PREHOOK: Output: default@dest_l1 -POSTHOOK: query: INSERT OVERWRITE TABLE dest_l1 -SELECT j.* -FROM (SELECT t1.key, p1.value - FROM src1 t1 - LEFT OUTER JOIN src p1 - ON (t1.key = p1.key) - UNION ALL - SELECT t2.key, p2.value - FROM src1 t2 - LEFT OUTER JOIN src p2 - ON (t2.key = p2.key)) j -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@src1 -POSTHOOK: Output: default@dest_l1 -POSTHOOK: Lineage: dest_l1.key EXPRESSION [(src1)t1.FieldSchema(name:key, type:string, comment:default), (src1)t2.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: dest_l1.value EXPRESSION [(src)p1.FieldSchema(name:value, type:string, comment:default), (src)p2.FieldSchema(name:value, type:string, comment:default), ] diff --git ql/src/test/results/clientpositive/list_bucket_dml_14.q.out ql/src/test/results/clientpositive/list_bucket_dml_14.q.out deleted file mode 100644 index 3e3e25a0cf..0000000000 --- ql/src/test/results/clientpositive/list_bucket_dml_14.q.out +++ /dev/null @@ -1,431 +0,0 @@ -PREHOOK: query: create table list_bucketing (key String, value String) -skewed by (key) on ("484") -stored as DIRECTORIES -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@list_bucketing -POSTHOOK: query: create table list_bucketing (key String, value String) -skewed by (key) on ("484") -stored as DIRECTORIES -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@list_bucketing -PREHOOK: query: explain extended -insert overwrite table list_bucketing select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@list_bucketing -POSTHOOK: query: explain extended -insert overwrite table list_bucketing select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@list_bucketing -OPTIMIZED SQL: SELECT `key`, `value` -FROM `default`.`src` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct list_bucketing { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct list_bucketing { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.list_bucketing - Is Table Level Stats: true - -PREHOOK: query: insert overwrite table list_bucketing select * from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@list_bucketing -POSTHOOK: query: insert overwrite table list_bucketing select * from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@list_bucketing -POSTHOOK: Lineage: list_bucketing.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: list_bucketing.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: desc formatted list_bucketing -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@list_bucketing -POSTHOOK: query: desc formatted list_bucketing -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@list_bucketing -# col_name data_type comment -key string -value string - -# Detailed Table Information -Database: default -#### A masked pattern was here #### -Retention: 0 -#### A masked pattern was here #### -Table Type: MANAGED_TABLE -Table Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - bucketing_version 2 - numFiles 2 - numRows 500 - rawDataSize 5312 - totalSize 5812 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -InputFormat: org.apache.hadoop.mapred.TextInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Stored As SubDirectories: Yes -Skewed Columns: [key] -Skewed Values: [[484]] -#### A masked pattern was here #### -Skewed Value to Truncated Path: {[484]=/list_bucketing/key=484} -Storage Desc Params: - serialization.format 1 -PREHOOK: query: select count(1) from src -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from src -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -500 -PREHOOK: query: select count(1) from list_bucketing -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from list_bucketing -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -500 -PREHOOK: query: select key, value from src where key = "484" -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select key, value from src where key = "484" -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -484 val_484 -PREHOOK: query: explain extended -select key, value from list_bucketing where key = "484" -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -POSTHOOK: query: explain extended -select key, value from list_bucketing where key = "484" -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, `value` -FROM `default`.`list_bucketing` -WHERE `key` = '484' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing - filterExpr: (key = '484') (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '484') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: list_bucketing - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing - numFiles 2 - numRows 500 - rawDataSize 5312 - serialization.ddl struct list_bucketing { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing - numFiles 2 - numRows 500 - rawDataSize 5312 - serialization.ddl struct list_bucketing { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing - name: default.list_bucketing - Truncated Path -> Alias: - /list_bucketing [list_bucketing] - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key, value from list_bucketing where key = "484" -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -POSTHOOK: query: select key, value from list_bucketing where key = "484" -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing -#### A masked pattern was here #### -484 val_484 -PREHOOK: query: drop table list_bucketing -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@list_bucketing -PREHOOK: Output: default@list_bucketing -POSTHOOK: query: drop table list_bucketing -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@list_bucketing -POSTHOOK: Output: default@list_bucketing diff --git ql/src/test/results/clientpositive/list_bucket_dml_4.q.out ql/src/test/results/clientpositive/list_bucket_dml_4.q.out deleted file mode 100644 index a0334a9556..0000000000 --- ql/src/test/results/clientpositive/list_bucket_dml_4.q.out +++ /dev/null @@ -1,984 +0,0 @@ -PREHOOK: query: create table list_bucketing_static_part_n2 (key String, value String) - partitioned by (ds String, hr String) - skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103')) - stored as DIRECTORIES - STORED AS RCFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@list_bucketing_static_part_n2 -POSTHOOK: query: create table list_bucketing_static_part_n2 (key String, value String) - partitioned by (ds String, hr String) - skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103')) - stored as DIRECTORIES - STORED AS RCFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@list_bucketing_static_part_n2 -PREHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -OPTIMIZED SQL: SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008-04-08 - hr 11 - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.list_bucketing_static_part_n2 - Is Table Level Stats: false - -PREHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: show partitions list_bucketing_static_part_n2 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: query: show partitions list_bucketing_static_part_n2 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@list_bucketing_static_part_n2 -ds=2008-04-08/hr=11 -PREHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n2 -# col_name data_type comment -key string -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 11] -Database: default -Table: list_bucketing_static_part_n2 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 6 - numRows 1000 - rawDataSize 9624 - totalSize 10898 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Stored As SubDirectories: Yes -Skewed Columns: [key, value] -Skewed Values: [[103, val_103], [484, val_484], [51, val_14]] -#### A masked pattern was here #### -Skewed Value to Truncated Path: {[103, val_103]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=103/value=val_103, [484, val_484]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=484/value=val_484} -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -OPTIMIZED SQL: SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008-04-08 - hr 11 - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.list_bucketing_static_part_n2 - Is Table Level Stats: false - - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - name: default.list_bucketing_static_part_n2 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - name: default.list_bucketing_static_part_n2 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: show partitions list_bucketing_static_part_n2 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: query: show partitions list_bucketing_static_part_n2 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@list_bucketing_static_part_n2 -ds=2008-04-08/hr=11 -PREHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n2 -# col_name data_type comment -key string -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 11] -Database: default -Table: list_bucketing_static_part_n2 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 4 - numRows 1000 - rawDataSize 9624 - totalSize 10786 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Stored As SubDirectories: Yes -Skewed Columns: [key, value] -Skewed Values: [[103, val_103], [484, val_484], [51, val_14]] -#### A masked pattern was here #### -Skewed Value to Truncated Path: {[103, val_103]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=103/value=val_103, [484, val_484]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=484/value=val_484} -Storage Desc Params: - serialization.format 1 -PREHOOK: query: select count(1) from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -#### A masked pattern was here #### -1000 -PREHOOK: query: select count(*) from list_bucketing_static_part_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n2 -#### A masked pattern was here #### -POSTHOOK: query: select count(*) from list_bucketing_static_part_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n2 -#### A masked pattern was here #### -1000 -PREHOOK: query: explain extended -select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n2 -PREHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: explain extended -select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) AS `$f1`, CAST('2008-04-08' AS STRING) AS `$f2`, CAST('11' AS STRING) AS `$f3` -FROM `default`.`list_bucketing_static_part_n2` -WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '2008-04-08' AND `hr` = '11' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_static_part_n2 - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: value=val_484 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - numFiles 4 - numRows 1000 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 9624 - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10786 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n2 - name: default.list_bucketing_static_part_n2 - Truncated Path -> Alias: - /list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=484/value=val_484 [list_bucketing_static_part_n2] - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n2 -PREHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -484 val_484 2008-04-08 11 -484 val_484 2008-04-08 11 -PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -484 val_484 2008-04-08 11 -484 val_484 2008-04-08 12 -PREHOOK: query: drop table list_bucketing_static_part_n2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@list_bucketing_static_part_n2 -PREHOOK: Output: default@list_bucketing_static_part_n2 -POSTHOOK: query: drop table list_bucketing_static_part_n2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n2 -POSTHOOK: Output: default@list_bucketing_static_part_n2 diff --git ql/src/test/results/clientpositive/list_bucket_dml_9.q.out ql/src/test/results/clientpositive/list_bucket_dml_9.q.out deleted file mode 100644 index c16fa5d41a..0000000000 --- ql/src/test/results/clientpositive/list_bucket_dml_9.q.out +++ /dev/null @@ -1,984 +0,0 @@ -PREHOOK: query: create table list_bucketing_static_part_n0 (key String, value String) - partitioned by (ds String, hr String) - skewed by (key) on ('484','103') - stored as DIRECTORIES - STORED AS RCFILE -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@list_bucketing_static_part_n0 -POSTHOOK: query: create table list_bucketing_static_part_n0 (key String, value String) - partitioned by (ds String, hr String) - skewed by (key) on ('484','103') - stored as DIRECTORIES - STORED AS RCFILE -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@list_bucketing_static_part_n0 -PREHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -OPTIMIZED SQL: SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008-04-08 - hr 11 - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.list_bucketing_static_part_n0 - Is Table Level Stats: false - -PREHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: show partitions list_bucketing_static_part_n0 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: query: show partitions list_bucketing_static_part_n0 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@list_bucketing_static_part_n0 -ds=2008-04-08/hr=11 -PREHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n0 -# col_name data_type comment -key string -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 11] -Database: default -Table: list_bucketing_static_part_n0 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 6 - numRows 1000 - rawDataSize 9624 - totalSize 10898 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Stored As SubDirectories: Yes -Skewed Columns: [key] -Skewed Values: [[103], [484]] -#### A masked pattern was here #### -Skewed Value to Truncated Path: {[103]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=103, [484]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=484} -Storage Desc Params: - serialization.format 1 -PREHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: query: explain extended -insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -OPTIMIZED SQL: SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `ds` = '2008-04-08' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2008-04-08 - hr 11 - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - - Stage: Stage-2 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.list_bucketing_static_part_n0 - Is Table Level Stats: false - - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - name: default.list_bucketing_static_part_n0 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - name: default.list_bucketing_static_part_n0 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - -PREHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') -select key, value from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: show partitions list_bucketing_static_part_n0 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: query: show partitions list_bucketing_static_part_n0 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@list_bucketing_static_part_n0 -ds=2008-04-08/hr=11 -PREHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n0 -# col_name data_type comment -key string -value string - -# Partition Information -# col_name data_type comment -ds string -hr string - -# Detailed Partition Information -Partition Value: [2008-04-08, 11] -Database: default -Table: list_bucketing_static_part_n0 -#### A masked pattern was here #### -Partition Parameters: - COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 4 - numRows 1000 - rawDataSize 9624 - totalSize 10786 -#### A masked pattern was here #### - -# Storage Information -SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat -OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat -Compressed: No -Num Buckets: -1 -Bucket Columns: [] -Sort Columns: [] -Stored As SubDirectories: Yes -Skewed Columns: [key] -Skewed Values: [[103], [484]] -#### A masked pattern was here #### -Skewed Value to Truncated Path: {[103]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=103, [484]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=484} -Storage Desc Params: - serialization.format 1 -PREHOOK: query: select count(1) from srcpart where ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -#### A masked pattern was here #### -POSTHOOK: query: select count(1) from srcpart where ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -#### A masked pattern was here #### -1000 -PREHOOK: query: select count(*) from list_bucketing_static_part_n0 -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n0 -#### A masked pattern was here #### -POSTHOOK: query: select count(*) from list_bucketing_static_part_n0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n0 -#### A masked pattern was here #### -1000 -PREHOOK: query: explain extended -select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n0 -PREHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: explain extended -select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) AS `$f1`, CAST('2008-04-08' AS STRING) AS `$f2`, CAST('11' AS STRING) AS `$f3` -FROM `default`.`list_bucketing_static_part_n0` -WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '2008-04-08' AND `hr` = '11' -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_static_part_n0 - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: key=484 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - numFiles 4 - numRows 1000 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 9624 - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10786 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_static_part_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n0 - name: default.list_bucketing_static_part_n0 - Truncated Path -> Alias: - /list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=484 [list_bucketing_static_part_n0] - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@list_bucketing_static_part_n0 -PREHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -POSTHOOK: query: select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 -#### A masked pattern was here #### -484 val_484 2008-04-08 11 -484 val_484 2008-04-08 11 -PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' -PREHOOK: type: QUERY -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -484 val_484 2008-04-08 11 -484 val_484 2008-04-08 12 -PREHOOK: query: drop table list_bucketing_static_part_n0 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@list_bucketing_static_part_n0 -PREHOOK: Output: default@list_bucketing_static_part_n0 -POSTHOOK: query: drop table list_bucketing_static_part_n0 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@list_bucketing_static_part_n0 -POSTHOOK: Output: default@list_bucketing_static_part_n0 diff --git ql/src/test/results/clientpositive/groupby9.q.out ql/src/test/results/clientpositive/llap/groupby9.q.out similarity index 55% rename from ql/src/test/results/clientpositive/groupby9.q.out rename to ql/src/test/results/clientpositive/llap/groupby9.q.out index 8eaa2e9d1f..0ef88e06ef 100644 --- ql/src/test/results/clientpositive/groupby9.q.out +++ ql/src/test/results/clientpositive/llap/groupby9.q.out @@ -32,91 +32,169 @@ POSTHOOK: Output: default@dest1_n117 POSTHOOK: Output: default@dest2_n31 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), value (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n117 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), value (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n117 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -128,7 +206,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n117 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: Column Stats Desc: @@ -136,85 +214,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n117 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.dest2_n31 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 Move Operator tables: @@ -225,29 +224,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31 - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.dest2_n31 PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST1_n117 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) GROUP BY SRC.key @@ -920,91 +903,169 @@ POSTHOOK: Output: default@dest1_n117 POSTHOOK: Output: default@dest2_n31 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: value (type: string), key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n117 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: value (type: string), key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n117 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -1016,7 +1077,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n117 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: Column Stats Desc: @@ -1024,85 +1085,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n117 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.dest2_n31 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 Move Operator tables: @@ -1113,29 +1095,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31 - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.dest2_n31 PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST1_n117 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) GROUP BY SRC.key @@ -1808,91 +1774,169 @@ POSTHOOK: Output: default@dest1_n117 POSTHOOK: Output: default@dest2_n31 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), value (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n117 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), value (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n117 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -1904,7 +1948,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n117 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: Column Stats Desc: @@ -1912,85 +1956,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n117 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.dest2_n31 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 Move Operator tables: @@ -2001,29 +1966,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31 - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.dest2_n31 PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST1_n117 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) GROUP BY SRC.key @@ -2696,92 +2645,171 @@ POSTHOOK: Output: default@dest1_n117 POSTHOOK: Output: default@dest2_n31 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(substr(value, 5)) + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(substr(value, 5)) + keys: key (type: string), value (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(substr(value, 5)) - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n117 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(substr(value, 5)) - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n117 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -2793,7 +2821,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n117 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: Column Stats Desc: @@ -2801,86 +2829,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n117 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.dest2_n31 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 Move Operator tables: @@ -2891,29 +2839,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31 - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.dest2_n31 PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST1_n117 SELECT SRC.key, COUNT(SUBSTR(SRC.value,5)) GROUP BY SRC.key @@ -3586,91 +3518,169 @@ POSTHOOK: Output: default@dest1_n117 POSTHOOK: Output: default@dest2_n31 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4, Stage-7 - Stage-4 depends on stages: Stage-2 - Stage-6 depends on stages: Stage-1, Stage-4, Stage-7 - Stage-5 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-5 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: value (type: string), key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 45250 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n117 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT substr(value, 5)) - keys: value (type: string), key (type: string), substr(value, 5) (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n117 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 250 Data size: 47000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -3682,7 +3692,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n117 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: Column Stats Desc: @@ -3690,85 +3700,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n117 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-6 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, val1, val2 - Column Types: int, string, string - Table: default.dest2_n31 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 Move Operator tables: @@ -3779,29 +3710,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31 - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.dest2_n31 PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST1_n117 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) GROUP BY SRC.key diff --git ql/src/test/results/clientpositive/groupby_complex_types.q.out ql/src/test/results/clientpositive/llap/groupby_complex_types.q.out similarity index 81% rename from ql/src/test/results/clientpositive/groupby_complex_types.q.out rename to ql/src/test/results/clientpositive/llap/groupby_complex_types.q.out index e784a5e04a..faa2c11b30 100644 --- ql/src/test/results/clientpositive/groupby_complex_types.q.out +++ ql/src/test/results/clientpositive/llap/groupby_complex_types.q.out @@ -44,89 +44,139 @@ POSTHOOK: Output: default@dest2_n41 POSTHOOK: Output: default@dest3_n7 STAGE DEPENDENCIES: Stage-3 is a root stage - Stage-0 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-0 - Stage-5 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-5 + Stage-4 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-4 Stage-6 depends on stages: Stage-1 - Stage-7 depends on stages: Stage-3 - Stage-2 depends on stages: Stage-7 - Stage-8 depends on stages: Stage-2 + Stage-2 depends on stages: Stage-4 + Stage-7 depends on stages: Stage-2 STAGE PLANS: Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + keys: array(key) (type: array) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: array) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: array) + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + keys: map(key:value) (type: map) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: map) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: map) + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + keys: struct(key,value) (type: struct) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: struct) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: struct) + Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(1) - keys: array(key) (type: array) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: array) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: array) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: array) + File Output Operator + compressed: false Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n163 + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(1) - keys: map(key:value) (type: map) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: map) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n41 + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(1) - keys: struct(key,value) (type: struct) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: struct) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: array) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n163 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest3_n7 + + Stage: Stage-4 + Dependency Collection Stage: Stage-0 Move Operator @@ -138,38 +188,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n163 - Stage: Stage-4 + Stage: Stage-5 Stats Work Basic Stats Work: - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: map) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: map) - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: map) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n41 - Stage: Stage-1 Move Operator tables: @@ -184,34 +206,6 @@ STAGE PLANS: Stats Work Basic Stats Work: - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: struct) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: struct) - Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: struct) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 106000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest3_n7 - Stage: Stage-2 Move Operator tables: @@ -222,7 +216,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest3_n7 - Stage: Stage-8 + Stage: Stage-7 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/llap/groupby_complex_types_multi_single_reducer.q.out ql/src/test/results/clientpositive/llap/groupby_complex_types_multi_single_reducer.q.out new file mode 100644 index 0000000000..5cd5d78713 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_complex_types_multi_single_reducer.q.out @@ -0,0 +1,262 @@ +PREHOOK: query: CREATE TABLE DEST1_n47(key ARRAY, value BIGINT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST1_n47 +POSTHOOK: query: CREATE TABLE DEST1_n47(key ARRAY, value BIGINT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST1_n47 +PREHOOK: query: CREATE TABLE DEST2_n8(key MAP, value BIGINT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST2_n8 +POSTHOOK: query: CREATE TABLE DEST2_n8(key MAP, value BIGINT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST2_n8 +PREHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 +INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n47 +PREHOOK: Output: default@dest2_n8 +POSTHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 +INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n47 +POSTHOOK: Output: default@dest2_n8 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: array(key) (type: array) + null sort order: z + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Group By Operator + aggregations: count(1) + keys: array(key) (type: array) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: array) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: array) + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: map(key:value) (type: map) + null sort order: z + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Group By Operator + aggregations: count(1) + keys: map(key:value) (type: map) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: map) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: map) + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: array) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: array) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: array), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 482000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 19280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 19280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n47 + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: map) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: map) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: map), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 232000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 9280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 9280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n8 + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n47 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n8 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + +PREHOOK: query: FROM SRC +INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 +INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n47 +PREHOOK: Output: default@dest2_n8 +POSTHOOK: query: FROM SRC +INSERT OVERWRITE TABLE DEST1_n47 SELECT ARRAY(SRC.key) as keyarray, COUNT(1) GROUP BY ARRAY(SRC.key) ORDER BY keyarray limit 10 +INSERT OVERWRITE TABLE DEST2_n8 SELECT MAP(SRC.key, SRC.value) as kvmap, COUNT(1) GROUP BY MAP(SRC.key, SRC.value) ORDER BY kvmap limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n47 +POSTHOOK: Output: default@dest2_n8 +POSTHOOK: Lineage: dest1_n47.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n47.value EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: dest2_n8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n8.value EXPRESSION [(src)src.null, ] +PREHOOK: query: SELECT DEST1_n47.* FROM DEST1_n47 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n47 +#### A masked pattern was here #### +POSTHOOK: query: SELECT DEST1_n47.* FROM DEST1_n47 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n47 +#### A masked pattern was here #### +["0"] 3 +["10"] 1 +["100"] 2 +["103"] 2 +["104"] 2 +["105"] 1 +["11"] 1 +["111"] 1 +["113"] 2 +["114"] 1 +PREHOOK: query: SELECT DEST2_n8.* FROM DEST2_n8 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n8 +#### A masked pattern was here #### +POSTHOOK: query: SELECT DEST2_n8.* FROM DEST2_n8 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n8 +#### A masked pattern was here #### +{"0":"val_0"} 3 +{"10":"val_10"} 1 +{"100":"val_100"} 2 +{"103":"val_103"} 2 +{"104":"val_104"} 2 +{"105":"val_105"} 1 +{"11":"val_11"} 1 +{"111":"val_111"} 1 +{"113":"val_113"} 2 +{"114":"val_114"} 1 diff --git ql/src/test/results/clientpositive/llap/groupby_cube1.q.out ql/src/test/results/clientpositive/llap/groupby_cube1.q.out new file mode 100644 index 0000000000..f166ee6784 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_cube1.q.out @@ -0,0 +1,895 @@ +PREHOOK: query: CREATE TABLE T1_n82(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n82 +POSTHOOK: query: CREATE TABLE T1_n82(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n82 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n82 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n82 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n82 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n82 +PREHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY CUBE(key, val) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY CUBE(key, val) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +1 11 1 +1 NULL 1 +2 12 1 +2 NULL 1 +3 13 1 +3 NULL 1 +7 17 1 +7 NULL 1 +8 18 1 +8 28 1 +8 NULL 2 +NULL 11 1 +NULL 12 1 +NULL 13 1 +NULL 17 1 +NULL 18 1 +NULL 28 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val, GROUPING__ID, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +1 11 0 1 +1 NULL 1 1 +2 12 0 1 +2 NULL 1 1 +3 13 0 1 +3 NULL 1 1 +7 17 0 1 +7 NULL 1 1 +8 18 0 1 +8 28 0 1 +8 NULL 1 2 +NULL 11 2 1 +NULL 12 2 1 +NULL 13 2 1 +NULL 17 2 1 +NULL 18 2 1 +NULL 28 2 1 +NULL NULL 3 6 +PREHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(DISTINCT val) + keys: key (type: string), 0L (type: bigint), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col2 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val, count(1) FROM T1_n82 GROUP BY key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +1 11 1 +1 NULL 1 +2 12 1 +2 NULL 1 +3 13 1 +3 NULL 1 +7 17 1 +7 NULL 1 +8 18 1 +8 28 1 +8 NULL 2 +NULL 11 1 +NULL 12 1 +NULL 13 1 +NULL 17 1 +NULL 18 1 +NULL 28 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(DISTINCT val) + keys: key (type: string), 0L (type: bigint), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: final + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col2 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n82 GROUP BY key with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +NULL 6 +PREHOOK: query: CREATE TABLE T2_n51(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n51 +POSTHOOK: query: CREATE TABLE T2_n51(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n51 +PREHOOK: query: CREATE TABLE T3_n16(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n16 +POSTHOOK: query: CREATE TABLE T3_n16(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n16 +PREHOOK: query: EXPLAIN +FROM T1_n82 +INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube +INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +PREHOOK: Output: default@t2_n51 +PREHOOK: Output: default@t3_n16 +POSTHOOK: query: EXPLAIN +FROM T1_n82 +INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube +INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +POSTHOOK: Output: default@t2_n51 +POSTHOOK: Output: default@t3_n16 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 1 (SIMPLE_EDGE) + Reducer 6 <- Reducer 5 (SIMPLE_EDGE) + Reducer 7 <- Reducer 6 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n82 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(1) + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(1) + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n51 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, val + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t3_n16 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, val + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 7 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n51 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key1, key2, val + Column Types: string, string, int + Table: default.t2_n51 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t3_n16 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key1, key2, val + Column Types: string, string, int + Table: default.t3_n16 + +PREHOOK: query: FROM T1_n82 +INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube +INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n82 +PREHOOK: Output: default@t2_n51 +PREHOOK: Output: default@t3_n16 +POSTHOOK: query: FROM T1_n82 +INSERT OVERWRITE TABLE T2_n51 SELECT key, val, count(1) group by key, val with cube +INSERT OVERWRITE TABLE T3_n16 SELECT key, val, sum(1) group by key, val with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n82 +POSTHOOK: Output: default@t2_n51 +POSTHOOK: Output: default@t3_n16 +POSTHOOK: Lineage: t2_n51.key1 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n51.key2 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n51.val EXPRESSION [(t1_n82)t1_n82.null, ] +POSTHOOK: Lineage: t3_n16.key1 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t3_n16.key2 SIMPLE [(t1_n82)t1_n82.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: t3_n16.val EXPRESSION [(t1_n82)t1_n82.null, ] diff --git ql/src/test/results/clientpositive/llap/groupby_cube_multi_gby.q.out ql/src/test/results/clientpositive/llap/groupby_cube_multi_gby.q.out new file mode 100644 index 0000000000..e742242791 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_cube_multi_gby.q.out @@ -0,0 +1,233 @@ +PREHOOK: query: create table t1_n21 like src +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t1_n21 +POSTHOOK: query: create table t1_n21 like src +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t1_n21 +PREHOOK: query: create table t2_n13 like src +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t2_n13 +POSTHOOK: query: create table t2_n13 like src +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t2_n13 +PREHOOK: query: explain from src +insert into table t1_n21 select +key, GROUPING__ID +group by cube(key, value) +insert into table t2_n13 select +key, value +group by key, value grouping sets ((key), (key, value)) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@t1_n21 +PREHOOK: Output: default@t2_n13 +POSTHOOK: query: explain from src +insert into table t1_n21 select +key, GROUPING__ID +group by cube(key, value) +insert into table t2_n13 select +key, value +group by key, value grouping sets ((key), (key, value)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@t1_n21 +POSTHOOK: Output: default@t2_n13 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), value (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), value (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 186000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n21 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 1000 Data size: 271000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n13 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n21 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.t1_n21 + + Stage: Stage-1 + Move Operator + tables: + replace: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n13 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.t2_n13 + diff --git ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out ql/src/test/results/clientpositive/llap/groupby_distinct_samekey.q.out similarity index 50% rename from ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out rename to ql/src/test/results/clientpositive/llap/groupby_distinct_samekey.q.out index 901d6378ff..4d6d6cf2ac 100644 --- ql/src/test/results/clientpositive/groupby_distinct_samekey.q.out +++ ql/src/test/results/clientpositive/llap/groupby_distinct_samekey.q.out @@ -32,43 +32,52 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n60 - Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: int1_n60 (type: int) - outputColumnNames: int1_n60 - Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n60 + Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: int1_n60 (type: int) + outputColumnNames: int1_n60 + Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(DISTINCT int1_n60) + keys: int1_n60 (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: sum(DISTINCT int1_n60) - keys: int1_n60 (type: int) - minReductionHashAggr: 0.99 - mode: hash + aggregations: sum(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int) + File Output Operator + compressed: false Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -90,43 +99,52 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n60 - Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: int1_n60 (type: int) - outputColumnNames: int1_n60 - Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n60 + Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: int1_n60 (type: int) + outputColumnNames: int1_n60 + Statistics: Num rows: 8 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(DISTINCT int1_n60) + keys: int1_n60 (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: sum(DISTINCT int1_n60) - keys: int1_n60 (type: int) - minReductionHashAggr: 0.99 - mode: hash + aggregations: sum(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int) + File Output Operator + compressed: false Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/groupby_duplicate_key.q.out ql/src/test/results/clientpositive/llap/groupby_duplicate_key.q.out new file mode 100644 index 0000000000..eba89d8da5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_duplicate_key.q.out @@ -0,0 +1,372 @@ +PREHOOK: query: explain +select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Row Limit Per Split: 10 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), '' (type: string), '' (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), '' (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), '' (type: string) + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), '' (type: string), '' (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), '' (type: string), '' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 63750 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select distinct key, "" as dummy1, "" as dummy2 from src tablesample (10 rows) order by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +165 +238 +255 +27 +278 +311 +409 +484 +86 +98 +PREHOOK: query: explain +create table dummy_n6 as +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy_n6 +POSTHOOK: query: explain +create table dummy_n6 as +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dummy_n6 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-0, Stage-2 + Stage-3 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Row Limit Per Split: 10 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), 'X' (type: string), 'X' (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), 'X' (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), 'X' (type: string) + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), 'X' (type: string), 'X' (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), 'X' (type: string), 'X' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dummy_n6 + Select Operator + expressions: _col0 (type: string), 'X' (type: string), 'X' (type: string) + outputColumnNames: col1, col2, col3 + Statistics: Num rows: 250 Data size: 64250 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll') + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-4 + Create Table + columns: key string, dummy1 string, dummy2 string + name: default.dummy_n6 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat + serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, dummy1, dummy2 + Column Types: string, string, string + Table: default.dummy_n6 + + Stage: Stage-0 + Move Operator + files: + hdfs directory: true +#### A masked pattern was here #### + +PREHOOK: query: create table dummy_n6 as +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@dummy_n6 +POSTHOOK: query: create table dummy_n6 as +select distinct key, "X" as dummy1, "X" as dummy2 from src tablesample (10 rows) order by key +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dummy_n6 +POSTHOOK: Lineage: dummy_n6.dummy1 SIMPLE [] +POSTHOOK: Lineage: dummy_n6.dummy2 SIMPLE [] +POSTHOOK: Lineage: dummy_n6.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select key,dummy1,dummy2 from dummy_n6 order by key +PREHOOK: type: QUERY +PREHOOK: Input: default@dummy_n6 +#### A masked pattern was here #### +POSTHOOK: query: select key,dummy1,dummy2 from dummy_n6 order by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dummy_n6 +#### A masked pattern was here #### +165 X X +238 X X +255 X X +27 X X +278 X X +311 X X +409 X X +484 X X +86 X X +98 X X +PREHOOK: query: explain +select max('pants'), max('pANTS') from src group by key order by key limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select max('pants'), max('pANTS') from src group by key order by key limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: key (type: string) + null sort order: z + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 1 + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: max('pants'), max('pANTS') + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: string), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: max(VALUE._col0), max(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string), _col2 (type: string), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 113750 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 455 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 455 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select max('pants'), max('pANTS') from src group by key order by key limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select max('pants'), max('pANTS') from src group by key order by key limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +pants pANTS diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_id3.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_id3.q.out new file mode 100644 index 0000000000..7e820d084d --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_id3.q.out @@ -0,0 +1,227 @@ +PREHOOK: query: CREATE TABLE T1_n86(key INT, value INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n86 +POSTHOOK: query: CREATE TABLE T1_n86(key INT, value INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n86 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n86 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n86 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n86 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n86 +PREHOOK: query: EXPLAIN +SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n86 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (_col2 = 1L) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), 1L (type: bigint) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: int), KEY._col1 (type: int), 1L (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +1 NULL 1 2 +2 NULL 1 1 +3 NULL 1 2 +4 NULL 1 1 +PREHOOK: query: EXPLAIN +SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n86 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (_col2 = 1L) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), 1L (type: bigint), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, value, GROUPING__ID, count(*) +FROM T1_n86 +GROUP BY key, value +GROUPING SETS ((), (key)) +HAVING GROUPING__ID = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n86 +#### A masked pattern was here #### +1 NULL 1 2 +2 NULL 1 1 +3 NULL 1 2 +4 NULL 1 1 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets1.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets1.q.out new file mode 100644 index 0000000000..97eea5572a --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets1.q.out @@ -0,0 +1,690 @@ +PREHOOK: query: CREATE TABLE T1_n41(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n41 +POSTHOOK: query: CREATE TABLE T1_n41(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n41 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n41 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n41 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n41 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n41 +PREHOOK: query: SELECT * FROM T1_n41 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM T1_n41 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +t1_n41.a t1_n41.b t1_n41.c +1 1 3 +2 2 4 +2 3 5 +3 2 8 +5 2 2 +8 1 1 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n41 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n41 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) from T1_n41 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) from T1_n41 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a b _c2 +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n41 group by cube(a, b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n41 group by cube(a, b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) from T1_n41 group by cube(a, b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) from T1_n41 group by cube(a, b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a b _c2 +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a b _c2 +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM T1_n41 GROUP BY a, b GROUPING SETS (a, (a, b)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a b _c2 +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +PREHOOK: query: EXPLAIN +SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: a, b, c + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string), c (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + null sort order: zzzz + sort order: ++++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a FROM T1_n41 GROUP BY a, b, c GROUPING SETS (a, b, c) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a +1 +2 +3 +5 +8 +NULL +NULL +NULL +NULL +NULL +NULL +NULL +NULL +NULL +PREHOOK: query: EXPLAIN +SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a FROM T1_n41 GROUP BY a GROUPING SETS ((a), (a)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +a +1 +2 +3 +5 +8 +PREHOOK: query: EXPLAIN +SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n41 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: (UDFToDouble(a) + UDFToDouble(b)) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a + b, count(*) FROM T1_n41 GROUP BY a + b GROUPING SETS (a+b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n41 +#### A masked pattern was here #### +_c0 _c1 +2.0 1 +4.0 1 +5.0 2 +7.0 1 +9.0 1 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets2.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets2.q.out new file mode 100644 index 0000000000..5a53460a83 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets2.q.out @@ -0,0 +1,483 @@ +PREHOOK: query: CREATE TABLE T1_n81(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n81 +POSTHOOK: query: CREATE TABLE T1_n81(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n81 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n81 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n81 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n81 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n81 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n81 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n81 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n81 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n81 group by cube(a, b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) from T1_n81 group by cube(a, b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n81 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) from T1_n81 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) from T1_n81 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +a b _c2 +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT a, b, sum(c) from T1_n81 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, sum(c) from T1_n81 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n81 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: a, b, c + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(c) + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: double) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: double) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, sum(c) from T1_n81 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, sum(c) from T1_n81 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +#### A masked pattern was here #### +a b _c2 +1 1 3.0 +1 NULL 3.0 +2 2 4.0 +2 3 5.0 +2 NULL 9.0 +3 2 8.0 +3 NULL 8.0 +5 2 2.0 +5 NULL 2.0 +8 1 1.0 +8 NULL 1.0 +NULL 1 4.0 +NULL 2 14.0 +NULL 3 5.0 +NULL NULL 23.0 +PREHOOK: query: CREATE TABLE T2_n50(a STRING, b STRING, c int, d int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n50 +POSTHOOK: query: CREATE TABLE T2_n50(a STRING, b STRING, c int, d int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n50 +PREHOOK: query: INSERT OVERWRITE TABLE T2_n50 +SELECT a, b, c, c from T1_n81 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n81 +PREHOOK: Output: default@t2_n50 +POSTHOOK: query: INSERT OVERWRITE TABLE T2_n50 +SELECT a, b, c, c from T1_n81 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n81 +POSTHOOK: Output: default@t2_n50 +POSTHOOK: Lineage: t2_n50.a SIMPLE [(t1_n81)t1_n81.FieldSchema(name:a, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n50.b SIMPLE [(t1_n81)t1_n81.FieldSchema(name:b, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n50.c EXPRESSION [(t1_n81)t1_n81.FieldSchema(name:c, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n50.d EXPRESSION [(t1_n81)t1_n81.FieldSchema(name:c, type:string, comment:null), ] +_col0 _col1 _col2 _col3 +PREHOOK: query: EXPLAIN +SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n50 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n50 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n50 + Statistics: Num rows: 6 Data size: 1068 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: a (type: string), b (type: string), (c + d) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1068 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.16666669 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 12 Data size: 2232 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 12 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 12 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n50 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, sum(c+d) from T2_n50 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n50 +#### A masked pattern was here #### +a b _c2 +1 1 6 +1 NULL 6 +2 2 8 +2 3 10 +2 NULL 18 +3 2 16 +3 NULL 16 +5 2 4 +5 NULL 4 +8 1 2 +8 NULL 2 +NULL 1 8 +NULL 2 28 +NULL 3 10 +NULL NULL 46 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets3.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets3.q.out new file mode 100644 index 0000000000..94ea09e806 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets3.q.out @@ -0,0 +1,319 @@ +PREHOOK: query: CREATE TABLE T1_n118(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n118 +POSTHOOK: query: CREATE TABLE T1_n118(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n118 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets1.txt' INTO TABLE T1_n118 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n118 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets1.txt' INTO TABLE T1_n118 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n118 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets2.txt' INTO TABLE T1_n118 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n118 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets2.txt' INTO TABLE T1_n118 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n118 +PREHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n118 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: a, b, c + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(c), count(c), count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by cube(a, b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by cube(a, b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n118 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: a, b, c + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(c), count(c), count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +a b _c2 _c3 +1 1 3.0 2 +1 2 2.0 1 +1 NULL 2.6666666666666665 3 +2 2 5.333333333333333 3 +2 3 5.0 2 +2 NULL 5.2 5 +3 2 8.0 1 +3 NULL 8.0 1 +5 1 2.0 1 +5 NULL 2.0 1 +8 1 1.0 2 +8 NULL 1.0 2 +NULL 1 2.0 5 +NULL 2 5.2 5 +NULL 3 5.0 2 +NULL NULL 3.8333333333333335 12 +PREHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +Explain +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n118 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: a, b, c + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(c), count(c), count() + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: double), _col4 (type: bigint), _col5 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col3 / _col4) (type: double), _col5 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, avg(c), count(*) from T1_n118 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n118 +#### A masked pattern was here #### +a b _c2 _c3 +1 1 3.0 2 +1 2 2.0 1 +1 NULL 2.6666666666666665 3 +2 2 5.333333333333333 3 +2 3 5.0 2 +2 NULL 5.2 5 +3 2 8.0 1 +3 NULL 8.0 1 +5 1 2.0 1 +5 NULL 2.0 1 +8 1 1.0 2 +8 NULL 1.0 2 +NULL 1 2.0 5 +NULL 2 5.2 5 +NULL 3 5.0 2 +NULL NULL 3.8333333333333335 12 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets4.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets4.q.out new file mode 100644 index 0000000000..b5bdddb049 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets4.q.out @@ -0,0 +1,502 @@ +PREHOOK: query: CREATE TABLE T1_n143(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n143 +POSTHOOK: query: CREATE TABLE T1_n143(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n143 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n143 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n143 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n143 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n143 +PREHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n143 + filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq2 +on subq1.a = subq2.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by cube(a, b) ) subq2 +on subq1.a = subq2.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n143 + filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +1 1 1 1 1 1 +1 1 1 1 NULL 1 +1 NULL 1 1 1 1 +1 NULL 1 1 NULL 1 +2 2 1 2 2 1 +2 2 1 2 3 1 +2 2 1 2 NULL 2 +2 3 1 2 2 1 +2 3 1 2 3 1 +2 3 1 2 NULL 2 +2 NULL 2 2 2 1 +2 NULL 2 2 3 1 +2 NULL 2 2 NULL 2 +PREHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 5 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n143 + filterExpr: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (UDFToDouble(a) < 3.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: bigint) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq1 +join +(SELECT a, b, count(*) from T1_n143 where a < 3 group by a, b with cube) subq2 +on subq1.a = subq2.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n143 +#### A masked pattern was here #### +1 1 1 1 1 1 +1 1 1 1 NULL 1 +1 NULL 1 1 1 1 +1 NULL 1 1 NULL 1 +2 2 1 2 2 1 +2 2 1 2 3 1 +2 2 1 2 NULL 2 +2 3 1 2 2 1 +2 3 1 2 3 1 +2 3 1 2 NULL 2 +2 NULL 2 2 2 1 +2 NULL 2 2 3 1 +2 NULL 2 2 NULL 2 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets5.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets5.q.out new file mode 100644 index 0000000000..b24d967f19 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets5.q.out @@ -0,0 +1,377 @@ +PREHOOK: query: CREATE TABLE T1_n24(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n24 +POSTHOOK: query: CREATE TABLE T1_n24(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n24 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n24 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n24 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n24 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n24 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n24 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by cube(a, b) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by cube(a, b) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n24 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n24 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: a, b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), 0L (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM +(SELECT a, b, count(1) from T1_n24 group by a, b) subq1 group by a, b with cube +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n24 +#### A masked pattern was here #### +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +8 NULL 1 +NULL 1 2 +NULL 2 3 +NULL 3 1 +NULL NULL 6 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets6.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets6.q.out new file mode 100644 index 0000000000..4510a71a3f --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets6.q.out @@ -0,0 +1,186 @@ +PREHOOK: query: CREATE TABLE T1_n75(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n75 +POSTHOOK: query: CREATE TABLE T1_n75(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n75 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n75 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n75 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n75 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n75 +PREHOOK: query: EXPLAIN +SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n75 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (UDFToDouble(a) = 5.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +5 2 +5 NULL +PREHOOK: query: EXPLAIN +SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n75 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (UDFToDouble(a) = 5.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string), b (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b FROM +(SELECT a, b from T1_n75 group by a, b grouping sets ( (a,b),a )) res +WHERE res.a=5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n75 +#### A masked pattern was here #### +5 2 +5 NULL diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets_grouping.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets_grouping.q.out new file mode 100644 index 0000000000..8851bf85c1 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets_grouping.q.out @@ -0,0 +1,1540 @@ +PREHOOK: query: CREATE TABLE T1_n64(key INT, value INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n64 +POSTHOOK: query: CREATE TABLE T1_n64(key INT, value INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n64 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n64 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n64 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/groupby_groupingid.txt' INTO TABLE T1_n64 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n64 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 0 +1 NULL 0 0 0 +1 NULL 1 0 1 +2 2 0 0 0 +2 NULL 1 0 1 +3 3 0 0 0 +3 NULL 0 0 0 +3 NULL 1 0 1 +4 5 0 0 0 +4 NULL 1 0 1 +NULL NULL 3 1 1 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 0 +1 NULL 0 0 0 +1 NULL 1 0 1 +2 2 0 0 0 +2 NULL 1 0 1 +3 3 0 0 0 +3 NULL 0 0 0 +3 NULL 1 0 1 +4 5 0 0 0 +4 NULL 1 0 1 +NULL 1 2 1 0 +NULL 2 2 1 0 +NULL 3 2 1 0 +NULL 5 2 1 0 +NULL NULL 2 1 0 +NULL NULL 3 1 1 +PREHOOK: query: explain +select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (grouping(_col2, 1L) = 1L) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +NULL 1 +NULL 2 +NULL 3 +NULL 5 +NULL NULL +NULL NULL +PREHOOK: query: explain +select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((grouping(_col2, 1L) = 1L) or (grouping(_col2, 0L) = 1L)) (type: boolean) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1L) + grouping(_col2, 0L)) (type: bigint), CASE WHEN (((grouping(_col2, 1L) + grouping(_col2, 0L)) = 1L)) THEN (_col0) ELSE (null) END (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: bigint), _col3 (type: int) + null sort order: zz + sort order: -+ + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: int) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 NULL 1 +2 NULL 1 +3 NULL 1 +4 NULL 1 +NULL 1 1 +NULL 2 1 +NULL 3 1 +NULL 5 1 +NULL NULL 1 +NULL NULL 2 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 0 +1 NULL 0 0 0 +1 NULL 1 0 1 +2 2 0 0 0 +2 NULL 1 0 1 +3 3 0 0 0 +3 NULL 0 0 0 +3 NULL 1 0 1 +4 5 0 0 0 +4 NULL 1 0 1 +NULL NULL 3 1 1 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L) (type: bigint), grouping(_col2, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key), grouping(value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 0 +1 NULL 0 0 0 +1 NULL 1 0 1 +2 2 0 0 0 +2 NULL 1 0 1 +3 3 0 0 0 +3 NULL 0 0 0 +3 NULL 1 0 1 +4 5 0 0 0 +4 NULL 1 0 1 +NULL 1 2 1 0 +NULL 2 2 1 0 +NULL 3 2 1 0 +NULL 5 2 1 0 +NULL NULL 2 1 0 +NULL NULL 3 1 1 +PREHOOK: query: explain +select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (grouping(_col2, 1L) = 1) (type: boolean) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +NULL 1 +NULL 2 +NULL 3 +NULL 5 +NULL NULL +NULL NULL +PREHOOK: query: explain +select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((grouping(_col2, 1L) = 1) or (grouping(_col2, 0L) = 1)) (type: boolean) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), (grouping(_col2, 1L) + grouping(_col2, 0L)) (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: bigint), CASE WHEN ((_col2 = 1L)) THEN (_col0) END (type: int) + null sort order: zz + sort order: -+ + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: int) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, grouping(key)+grouping(value) as x +from T1_n64 +group by cube(key, value) +having grouping(key) = 1 OR grouping(value) = 1 +order by x desc, case when x = 1 then key end +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 NULL 1 +2 NULL 1 +3 NULL 1 +4 NULL 1 +NULL 1 1 +NULL 2 1 +NULL 3 1 +NULL 5 1 +NULL NULL 1 +NULL NULL 2 +PREHOOK: query: explain +select key, value, grouping(key), grouping(value) +from T1_n64 +group by key, value +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, grouping(key), grouping(value) +from T1_n64 +group by key, value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), 0L (type: bigint), 0L (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, grouping(key), grouping(value) +from T1_n64 +group by key, value +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, grouping(key), grouping(value) +from T1_n64 +group by key, value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 +1 NULL 0 0 +2 2 0 0 +3 3 0 0 +3 NULL 0 0 +4 5 0 0 +PREHOOK: query: explain +select key, value, grouping(value) +from T1_n64 +group by key, value +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, grouping(value) +from T1_n64 +group by key, value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), 0L (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, grouping(value) +from T1_n64 +group by key, value +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, grouping(value) +from T1_n64 +group by key, value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 +1 NULL 0 +2 2 0 +3 3 0 +3 NULL 0 +4 5 0 +PREHOOK: query: explain +select key, value +from T1_n64 +group by key, value +having grouping(key) = 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value +from T1_n64 +group by key, value +having grouping(key) = 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value +from T1_n64 +group by key, value +having grouping(key) = 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value +from T1_n64 +group by key, value +having grouping(key) = 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 +1 NULL +2 2 +3 3 +3 NULL +4 5 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 +1 NULL 0 0 +1 NULL 1 1 +2 2 0 0 +2 NULL 1 1 +3 3 0 0 +3 NULL 0 0 +3 NULL 1 1 +4 5 0 0 +4 NULL 1 1 +NULL 1 2 2 +NULL 2 2 2 +NULL 3 2 2 +NULL 5 2 2 +NULL NULL 2 2 +NULL NULL 3 3 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 32 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0L, 1L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by cube(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by cube(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 +1 NULL 0 0 +1 NULL 1 2 +2 2 0 0 +2 NULL 1 2 +3 3 0 0 +3 NULL 0 0 +3 NULL 1 2 +4 5 0 0 +4 NULL 1 2 +NULL 1 2 1 +NULL 2 2 1 +NULL 3 2 1 +NULL 5 2 1 +NULL NULL 2 1 +NULL NULL 3 3 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 1L, 0L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(key, value) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 +1 NULL 0 0 +1 NULL 1 1 +2 2 0 0 +2 NULL 1 1 +3 3 0 0 +3 NULL 0 0 +3 NULL 1 1 +4 5 0 0 +4 NULL 1 1 +NULL NULL 3 3 +PREHOOK: query: explain +select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n64 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: int) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: int), value (type: int), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int), _col2 (type: bigint) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: bigint), grouping(_col2, 0L, 1L) (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by rollup(key, value) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +POSTHOOK: query: select key, value, `grouping__id`, grouping(value, key) +from T1_n64 +group by rollup(key, value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n64 +#### A masked pattern was here #### +1 1 0 0 +1 NULL 0 0 +1 NULL 1 2 +2 2 0 0 +2 NULL 1 2 +3 3 0 0 +3 NULL 0 0 +3 NULL 1 2 +4 5 0 0 +4 NULL 1 2 +NULL NULL 3 3 diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_sets_limit.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets_limit.q.out new file mode 100644 index 0000000000..08d8efa9ee --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_sets_limit.q.out @@ -0,0 +1,696 @@ +PREHOOK: query: CREATE TABLE T1_n141(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n141 +POSTHOOK: query: CREATE TABLE T1_n141(a STRING, b STRING, c STRING) ROW FORMAT DELIMITED FIELDS TERMINATED BY ' ' STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n141 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n141 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n141 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/grouping_sets.txt' INTO TABLE T1_n141 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n141 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: a (type: string), b (type: string) + null sort order: zz + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint), _col2 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + null sort order: zzz + sort order: +++ + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b WITH CUBE ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: a (type: string), b (type: string) + null sort order: zz + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 4 Data size: 1472 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint), _col2 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + null sort order: zzz + sort order: +++ + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b), b, ()) ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +PREHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: a (type: string), b (type: string) + null sort order: zz + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint), _col2 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + null sort order: zzz + sort order: +++ + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a, b, count(*) FROM T1_n141 GROUP BY a, b GROUPING SETS (a, (a, b)) ORDER BY a, b, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +1 1 1 +1 NULL 1 +2 2 1 +2 3 1 +2 NULL 2 +3 2 1 +3 NULL 1 +5 2 1 +5 NULL 1 +8 1 1 +PREHOOK: query: EXPLAIN +SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: +++ + keys: a (type: string), b (type: string), c (type: string) + null sort order: zzz + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: a (type: string), b (type: string), c (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + null sort order: zzzz + sort order: ++++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + Statistics: Num rows: 3 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: bigint) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a FROM T1_n141 GROUP BY a, b, c GROUPING SETS (a, b, c) ORDER BY a, b, c, GROUPING__ID LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +1 +2 +3 +5 +8 +NULL +NULL +NULL +NULL +NULL +PREHOOK: query: EXPLAIN +SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: + + keys: a (type: string) + null sort order: z + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: a (type: string) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: a (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a FROM T1_n141 GROUP BY a GROUPING SETS ((a), (a)) ORDER BY a LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +1 +2 +3 +5 +8 +PREHOOK: query: EXPLAIN +SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n141 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: + + keys: (UDFToDouble(a) + UDFToDouble(b)) (type: double) + null sort order: z + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: (UDFToDouble(a) + UDFToDouble(b)) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col1 (type: bigint), _col0 (type: double) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: double) + null sort order: z + sort order: + + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: double), VALUE._col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a + b, count(*) FROM T1_n141 GROUP BY a + b GROUPING SETS (a + b) ORDER BY a + b LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n141 +#### A masked pattern was here #### +2.0 1 +4.0 1 +5.0 2 +7.0 1 +9.0 1 diff --git ql/src/test/results/clientpositive/groupby_grouping_sets_view.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_sets_view.q.out similarity index 100% rename from ql/src/test/results/clientpositive/groupby_grouping_sets_view.q.out rename to ql/src/test/results/clientpositive/llap/groupby_grouping_sets_view.q.out diff --git ql/src/test/results/clientpositive/llap/groupby_grouping_window.q.out ql/src/test/results/clientpositive/llap/groupby_grouping_window.q.out new file mode 100644 index 0000000000..eda01a87b1 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_grouping_window.q.out @@ -0,0 +1,209 @@ +PREHOOK: query: create table t_n33(category int, live int, comments int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t_n33 +POSTHOOK: query: create table t_n33(category int, live int, comments int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t_n33 +PREHOOK: query: insert into table t_n33 select key, 0, 2 from src tablesample(3 rows) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@t_n33 +POSTHOOK: query: insert into table t_n33 select key, 0, 2 from src tablesample(3 rows) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@t_n33 +POSTHOOK: Lineage: t_n33.category EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: t_n33.comments SIMPLE [] +POSTHOOK: Lineage: t_n33.live SIMPLE [] +PREHOOK: query: explain +select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 +FROM t_n33 +GROUP BY category +GROUPING SETS ((), (category)) +HAVING max(comments) > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@t_n33 +#### A masked pattern was here #### +POSTHOOK: query: explain +select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 +FROM t_n33 +GROUP BY category +GROUPING SETS ((), (category)) +HAVING max(comments) > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_n33 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t_n33 + Statistics: Num rows: 3 Data size: 36 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: category (type: int), live (type: int), comments (type: int) + outputColumnNames: category, live, comments + Statistics: Num rows: 3 Data size: 36 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: max(live), max(comments) + keys: category (type: int), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: bigint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: bigint) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int), _col3 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: max(VALUE._col0), max(VALUE._col1) + keys: KEY._col0 (type: int), KEY._col1 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col2, _col3 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Filter Operator + predicate: (_col3 > 0) (type: boolean) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col3 (type: int) + null sort order: az + sort order: ++ + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: int), VALUE._col1 (type: int), KEY.reducesinkkey1 (type: int) + outputColumnNames: _col0, _col2, _col3 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + PTF Operator + Function definitions: + Input definition + input alias: ptf_0 + output shape: _col0: int, _col2: int, _col3: int + type: WINDOWING + Windowing table definition + input alias: ptf_1 + name: windowingtablefunction + order by: _col3 ASC NULLS LAST + partition by: _col0 + raw input shape: + window functions: + window function definition + alias: rank_window_0 + arguments: _col3 + name: rank + window function: GenericUDAFRankEvaluator + window frame: ROWS PRECEDING(MAX)~FOLLOWING(MAX) + isPivotResult: true + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), rank_window_0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 +FROM t_n33 +GROUP BY category +GROUPING SETS ((), (category)) +HAVING max(comments) > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@t_n33 +#### A masked pattern was here #### +POSTHOOK: query: select category, max(live) live, max(comments) comments, rank() OVER (PARTITION BY category ORDER BY comments) rank1 +FROM t_n33 +GROUP BY category +GROUPING SETS ((), (category)) +HAVING max(comments) > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_n33 +#### A masked pattern was here #### +NULL 0 2 1 +86 0 2 1 +238 0 2 1 +311 0 2 1 +PREHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +GROUPING SETS ((), (category)) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_n33 +#### A masked pattern was here #### +POSTHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +GROUPING SETS ((), (category)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_n33 +#### A masked pattern was here #### +0 NULL +0 NULL +0 NULL +1 NULL +PREHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +PREHOOK: type: QUERY +PREHOOK: Input: default@t_n33 +#### A masked pattern was here #### +POSTHOOK: query: SELECT grouping(category), lead(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_n33 +#### A masked pattern was here #### +0 0 +0 0 +0 NULL +PREHOOK: query: SELECT grouping(category), lag(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +PREHOOK: type: QUERY +PREHOOK: Input: default@t_n33 +#### A masked pattern was here #### +POSTHOOK: query: SELECT grouping(category), lag(live) over(partition by grouping(category)) +FROM t_n33 +GROUP BY category, live +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_n33 +#### A masked pattern was here #### +0 NULL +0 0 +0 0 diff --git ql/src/test/results/clientpositive/llap/groupby_join_pushdown.q.out ql/src/test/results/clientpositive/llap/groupby_join_pushdown.q.out new file mode 100644 index 0000000000..8cdf422a63 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_join_pushdown.q.out @@ -0,0 +1,1606 @@ +PREHOOK: query: EXPLAIN +SELECT f.key, g.key, count(g.key) +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT f.key, g.key, count(g.key) +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col0) + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: string), (_col1 * _col3) (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT f.key, g.key +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT f.key, g.key +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT DISTINCT f.value, g.value +FROM src f JOIN src g ON(f.value = g.value) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT DISTINCT f.value, g.value +FROM src f JOIN src g ON(f.value = g.value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 22750 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT f.key, g.key, COUNT(*) +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT f.key, g.key, COUNT(*) +FROM src f JOIN src g ON(f.key = g.key) +GROUP BY f.key, g.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: string), (_col1 * _col3) (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT f.ctinyint, g.ctinyint, SUM(f.cbigint) +FROM alltypesorc f JOIN alltypesorc g ON(f.cint = g.cint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT f.ctinyint, g.ctinyint, SUM(f.cbigint) +FROM alltypesorc f JOIN alltypesorc g ON(f.cint = g.cint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: cint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 146784 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int), cbigint (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 9173 Data size: 109584 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: tinyint), _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 4 + Map Operator Tree: + TableScan + alias: g + filterExpr: cint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: tinyint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: int) + 1 _col1 (type: int) + outputColumnNames: _col0, _col2, _col3 + Statistics: Num rows: 13785 Data size: 183376 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2) + keys: _col0 (type: tinyint), _col3 (type: tinyint) + minReductionHashAggr: 0.8912586 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint), _col1 (type: tinyint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) + Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6892 Data size: 100984 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN +SELECT f.cbigint, g.cbigint, MAX(f.cint) +FROM alltypesorc f JOIN alltypesorc g ON(f.cbigint = g.cbigint) +GROUP BY f.cbigint, g.cbigint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT f.cbigint, g.cbigint, MAX(f.cint) +FROM alltypesorc f JOIN alltypesorc g ON(f.cbigint = g.cbigint) +GROUP BY f.cbigint, g.cbigint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: cbigint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 110088 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cbigint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 82188 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: cint (type: int), cbigint (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 82188 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: max(_col0) + keys: _col1 (type: bigint) + minReductionHashAggr: 0.49994552 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: bigint) + Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: cbigint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cbigint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: cbigint (type: bigint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: bigint) + minReductionHashAggr: 0.49994552 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: bigint) + Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: bigint) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 4586 Data size: 27400 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: max(VALUE._col0) + keys: KEY._col0 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4586 Data size: 45744 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: bigint) + 1 _col0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: bigint), _col2 (type: bigint), _col1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 4586 Data size: 73144 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT f.ctinyint, g.ctinyint, MIN(f.ctinyint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT f.ctinyint, g.ctinyint, MIN(f.ctinyint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: min(_col0) + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: tinyint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: min(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: tinyint), _col2 (type: tinyint), _col1 (type: tinyint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 132 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT MIN(f.cint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT MIN(f.cint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: min(_col1) + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 131 Data size: 396 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: min(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col1 + Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 132 Data size: 528 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT count(f.ctinyint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT count(f.ctinyint) +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col0) + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col1, _col3 + Statistics: Num rows: 132 Data size: 2112 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: (_col1 * _col3) (type: bigint) + outputColumnNames: _col0 + Statistics: Num rows: 132 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 132 Data size: 1056 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT count(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT count(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col1) + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 132 Data size: 2512 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: (_col1 * _col3) (type: bigint), _col0 (type: tinyint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 3 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: tinyint) + minReductionHashAggr: 0.98571897 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 131 Data size: 1444 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 132 Data size: 2512 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: (_col1 * _col3) (type: bigint), _col0 (type: tinyint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 132 Data size: 1456 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT pk_alltypesorc_1 PRIMARY KEY (ctinyint) DISABLE RELY +PREHOOK: type: ALTERTABLE_ADDCONSTRAINT +POSTHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT pk_alltypesorc_1 PRIMARY KEY (ctinyint) DISABLE RELY +POSTHOOK: type: ALTERTABLE_ADDCONSTRAINT +PREHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 4 + Map Operator Tree: + TableScan + alias: g + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1161499 Data size: 13900620 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: tinyint), _col2 (type: tinyint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint), _col1 (type: tinyint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) + Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 17161 Data size: 274216 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: bigint), _col0 (type: tinyint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 17161 Data size: 205752 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 17161 Data size: 205752 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: ALTER TABLE alltypesorc DROP CONSTRAINT pk_alltypesorc_1 +PREHOOK: type: ALTERTABLE_DROPCONSTRAINT +POSTHOOK: query: ALTER TABLE alltypesorc DROP CONSTRAINT pk_alltypesorc_1 +POSTHOOK: type: ALTERTABLE_DROPCONSTRAINT +PREHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT uk_alltypesorc_1 UNIQUE (ctinyint) DISABLE RELY +PREHOOK: type: ALTERTABLE_ADDCONSTRAINT +POSTHOOK: query: ALTER TABLE alltypesorc ADD CONSTRAINT uk_alltypesorc_1 UNIQUE (ctinyint) DISABLE RELY +POSTHOOK: type: ALTERTABLE_ADDCONSTRAINT +PREHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +PREHOOK: type: QUERY +PREHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT sum(f.cint), f.ctinyint +FROM alltypesorc f JOIN alltypesorc g ON(f.ctinyint = g.ctinyint) +GROUP BY f.ctinyint, g.ctinyint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@alltypesorc +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: f + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 73392 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint), cint (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 9173 Data size: 54792 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: all inputs + Map 4 + Map Operator Tree: + TableScan + alias: g + filterExpr: ctinyint is not null (type: boolean) + Statistics: Num rows: 12288 Data size: 36696 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ctinyint is not null (type: boolean) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ctinyint (type: tinyint) + outputColumnNames: _col0 + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: tinyint) + Statistics: Num rows: 9173 Data size: 27396 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: tinyint) + 1 _col0 (type: tinyint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 647260 Data size: 7739232 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: tinyint), _col2 (type: tinyint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: tinyint), _col1 (type: tinyint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: tinyint), _col1 (type: tinyint) + Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: tinyint), KEY._col1 (type: tinyint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 17161 Data size: 274088 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: bigint), _col0 (type: tinyint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 17161 Data size: 205688 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 17161 Data size: 205688 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out new file mode 100644 index 0000000000..09b9e2f37b --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out @@ -0,0 +1,368 @@ +PREHOOK: query: CREATE TABLE dest1_n144(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n144 +POSTHOOK: query: CREATE TABLE dest1_n144(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n144 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n144 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n144 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n144 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n144 +OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +GROUP BY SUBSTR(`key`, 1, 1) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT _col1), sum(_col1) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 93500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93500 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col3 (type: double) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [src] + /srcpart/ds=2008-04-08/hr=12 [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 31916 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2 + columns.comments + columns.types string:int:string +#### A masked pattern was here #### + name default.dest1_n144 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n144 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + outputColumnNames: key, c1, c2 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2 + columns.comments + columns.types string:int:string +#### A masked pattern was here #### + name default.dest1_n144 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n144 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, c1, c2 + Column Types: string, int, string + Table: default.dest1_n144 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n144 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n144 +POSTHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n144 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n144 +POSTHOOK: Lineage: dest1_n144.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n144.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n144.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n144.* FROM dest1_n144 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n144 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n144.* FROM dest1_n144 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n144 +#### A masked pattern was here #### +0 1 00.0 +1 71 132828.0 +2 69 251142.0 +3 62 364008.0 +4 74 4105526.0 +5 6 5794.0 +6 5 6796.0 +7 6 71470.0 +8 8 81524.0 +9 7 92094.0 diff --git ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out new file mode 100644 index 0000000000..c5830c769d --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out @@ -0,0 +1,370 @@ +PREHOOK: query: CREATE TABLE dest1_n174(key STRING, c1 INT, c2 STRING, C3 INT, c4 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n174 +POSTHOOK: query: CREATE TABLE dest1_n174(key STRING, c1 INT, c2 STRING, C3 INT, c4 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n174 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n174 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n174 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n174 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n174 +OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2`, SUM(DISTINCT SUBSTR(`value`, 5)) AS `_o__c3`, COUNT(DISTINCT `value`) AS `_o__c4` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +GROUP BY SUBSTR(`key`, 1, 1) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT _col1), sum(_col1), sum(DISTINCT _col1), count(DISTINCT _col2) + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1000 Data size: 294000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + numBuckets: -1 + sort order: +++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1000 Data size: 294000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col4 (type: double) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [src] + /srcpart/ds=2008-04-08/hr=12 [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(VALUE._col1), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1_n174 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n174 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) + outputColumnNames: key, c1, c2, c3, c4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2152 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 2152 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types struct:struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1_n174 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n174 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, c1, c2, c3, c4 + Column Types: string, int, string, int, int + Table: default.dest1_n174 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n174 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n174 +POSTHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n174 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n174 +POSTHOOK: Lineage: dest1_n174.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n174.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n174.c3 EXPRESSION [(srcpart)src.null, ] +POSTHOOK: Lineage: dest1_n174.c4 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n174.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n174.* FROM dest1_n174 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n174 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n174.* FROM dest1_n174 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n174 +#### A masked pattern was here #### +0 1 00.0 0 1 +1 71 132828.0 10044 71 +2 69 251142.0 15780 69 +3 62 364008.0 20119 62 +4 74 4105526.0 30965 74 +5 6 5794.0 278 6 +6 5 6796.0 331 5 +7 6 71470.0 447 6 +8 8 81524.0 595 8 +9 7 92094.0 577 7 diff --git ql/src/test/results/clientpositive/llap/groupby_multi_insert_common_distinct.q.out ql/src/test/results/clientpositive/llap/groupby_multi_insert_common_distinct.q.out new file mode 100644 index 0000000000..aec09dec3c --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_multi_insert_common_distinct.q.out @@ -0,0 +1,280 @@ +PREHOOK: query: create table dest1_n99(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n99 +POSTHOOK: query: create table dest1_n99(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n99 +PREHOOK: query: create table dest2_n27(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest2_n27 +POSTHOOK: query: create table dest2_n27(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest2_n27 +PREHOOK: query: explain +from src +insert overwrite table dest1_n99 select key, count(distinct value) group by key +insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n99 +PREHOOK: Output: default@dest2_n27 +POSTHOOK: query: explain +from src +insert overwrite table dest1_n99 select key, count(distinct value) group by key +insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n99 +POSTHOOK: Output: default@dest2_n27 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT value) + keys: key (type: string), value (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT value) + keys: (key + key) (type: double), value (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: double), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n99 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n27 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n99 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest1_n99 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n27 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest2_n27 + +PREHOOK: query: from src +insert overwrite table dest1_n99 select key, count(distinct value) group by key +insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n99 +PREHOOK: Output: default@dest2_n27 +POSTHOOK: query: from src +insert overwrite table dest1_n99 select key, count(distinct value) group by key +insert overwrite table dest2_n27 select key+key, count(distinct value) group by key+key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n99 +POSTHOOK: Output: default@dest2_n27 +POSTHOOK: Lineage: dest1_n99.cnt EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n99.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n27.cnt EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n27.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from dest1_n99 where key < 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n99 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest1_n99 where key < 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n99 +#### A masked pattern was here #### +0 1 +2 1 +4 1 +5 1 +8 1 +9 1 +PREHOOK: query: select * from dest2_n27 where key < 20 order by key limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n27 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest2_n27 where key < 20 order by key limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n27 +#### A masked pattern was here #### +0 1 +10 1 +16 1 +18 1 +4 1 +8 1 diff --git ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer.q.out similarity index 57% rename from ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out rename to ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer.q.out index 756c179e8b..e95ba9776b 100644 --- ql/src/test/results/clientpositive/groupby_multi_single_reducer.q.out +++ ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer.q.out @@ -60,95 +60,108 @@ POSTHOOK: Output: default@dest_g3 POSTHOOK: Output: default@dest_g4 STAGE DEPENDENCIES: Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 - Stage-5 depends on stages: Stage-0 - Stage-1 depends on stages: Stage-3 - Stage-6 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-3 + Stage-2 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-4 + Stage-6 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-4 + Stage-7 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: substr(key, 1, 1) (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: substr(key, 1, 1) (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g4 - Filter Operator - predicate: (KEY._col0 >= 5) (type: boolean) - Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g2 - Filter Operator - predicate: (KEY._col0 < 5) (type: boolean) - Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g3 + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g4 + Filter Operator + predicate: (KEY._col0 >= 5) (type: boolean) + Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g2 + Filter Operator + predicate: (KEY._col0 < 5) (type: boolean) + Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g3 + + Stage: Stage-4 + Dependency Collection Stage: Stage-2 Move Operator @@ -160,7 +173,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g4 - Stage: Stage-4 + Stage: Stage-5 Stats Work Basic Stats Work: @@ -174,7 +187,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g2 - Stage: Stage-5 + Stage: Stage-6 Stats Work Basic Stats Work: @@ -188,7 +201,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g3 - Stage: Stage-6 + Stage: Stage-7 Stats Work Basic Stats Work: @@ -299,111 +312,202 @@ POSTHOOK: Output: default@dest_h2 POSTHOOK: Output: default@dest_h3 STAGE DEPENDENCIES: Stage-5 is a root stage - Stage-2 depends on stages: Stage-5 - Stage-6 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-5 - Stage-7 depends on stages: Stage-0 - Stage-1 depends on stages: Stage-5 - Stage-8 depends on stages: Stage-1 - Stage-9 depends on stages: Stage-5 - Stage-10 depends on stages: Stage-9 - Stage-3 depends on stages: Stage-10 + Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-6 + Stage-7 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-6 + Stage-8 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-6 + Stage-9 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-6 + Stage-10 depends on stages: Stage-4 + Stage-3 depends on stages: Stage-6 Stage-11 depends on stages: Stage-3 - Stage-4 depends on stages: Stage-9 - Stage-12 depends on stages: Stage-4 STAGE PLANS: Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: substr(key, 1, 1) (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: substr(key, 1, 1) (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: substr(key, 1, 1) (type: string) + null sort order: z + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: substr(key, 1, 1) (type: string), substr(key, 2, 1) (type: string), substr(value, 5) (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: substr(key, 1, 1) (type: string), substr(key, 2, 1) (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g4 - Filter Operator - predicate: (KEY._col0 >= 5) (type: boolean) - Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g2 - Filter Operator - predicate: (KEY._col0 < 5) (type: boolean) - Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g4 + Filter Operator + predicate: (KEY._col0 >= 5) (type: boolean) + Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g2 + Filter Operator + predicate: (KEY._col0 < 5) (type: boolean) + Statistics: Num rows: 166 Data size: 29216 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 12285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g3 + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (KEY._col0 >= 5) (type: boolean) + Statistics: Num rows: 166 Data size: 43326 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0), sum(KEY._col2:0._col0), count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 166 Data size: 32204 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col2) (type: int), concat(_col0, _col3) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 166 Data size: 46646 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 46646 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_h3 + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0), sum(KEY._col2:0._col0), count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col2 (type: bigint) + null sort order: zz + Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col2 (type: bigint), concat(_col0, _col3) (type: string), _col3 (type: double), _col4 (type: bigint) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint) + null sort order: zz + sort order: ++ + Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col2 (type: string), _col3 (type: double), _col4 (type: bigint) + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: bigint), VALUE._col0 (type: string), VALUE._col1 (type: double), VALUE._col2 (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 105 Data size: 29505 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_g3 + Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 2930 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), _col2 (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 10 Data size: 2810 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 2810 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_h2 + + Stage: Stage-6 + Dependency Collection Stage: Stage-2 Move Operator @@ -415,7 +519,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g4 - Stage: Stage-6 + Stage: Stage-7 Stats Work Basic Stats Work: @@ -429,7 +533,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g2 - Stage: Stage-7 + Stage: Stage-8 Stats Work Basic Stats Work: @@ -443,97 +547,11 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_g3 - Stage: Stage-8 + Stage: Stage-9 Stats Work Basic Stats Work: - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: substr(key, 1, 1) (type: string), substr(key, 2, 1) (type: string), substr(value, 5) (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: substr(key, 1, 1) (type: string), substr(key, 2, 1) (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Forward - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0), sum(KEY._col2:0._col0), count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint), concat(_col0, _col3) (type: string), _col3 (type: double), _col4 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (KEY._col0 >= 5) (type: boolean) - Statistics: Num rows: 166 Data size: 43326 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0), sum(KEY._col2:0._col0), count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 166 Data size: 32204 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col2) (type: int), concat(_col0, _col3) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 166 Data size: 46646 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 46646 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_h3 - - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: bigint) - null sort order: zz - sort order: ++ - Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col2 (type: string), _col3 (type: double), _col4 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: bigint), VALUE._col0 (type: string), VALUE._col1 (type: double), VALUE._col2 (type: bigint) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 146500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 10 - Statistics: Num rows: 10 Data size: 2930 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), _col2 (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 10 Data size: 2810 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 10 Data size: 2810 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_h2 - - Stage: Stage-3 + Stage: Stage-4 Move Operator tables: replace: true @@ -541,13 +559,13 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_h2 + name: default.dest_h3 - Stage: Stage-11 + Stage: Stage-10 Stats Work Basic Stats Work: - Stage: Stage-4 + Stage: Stage-3 Move Operator tables: replace: true @@ -555,9 +573,9 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_h3 + name: default.dest_h2 - Stage: Stage-12 + Stage: Stage-11 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer2.q.out ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer2.q.out new file mode 100644 index 0000000000..e52dafd8b8 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer2.q.out @@ -0,0 +1,276 @@ +PREHOOK: query: CREATE TABLE dest_g2_n4(key STRING, c1 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_g2_n4 +POSTHOOK: query: CREATE TABLE dest_g2_n4(key STRING, c1 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_g2_n4 +PREHOOK: query: CREATE TABLE dest_g3_n0(key STRING, c1 INT, c2 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_g3_n0 +POSTHOOK: query: CREATE TABLE dest_g3_n0(key STRING, c1 INT, c2 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_g3_n0 +PREHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) +INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest_g2_n4 +PREHOOK: Output: default@dest_g3_n0 +POSTHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) +INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest_g2_n4 +POSTHOOK: Output: default@dest_g3_n0 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: ((substr(key, 1, 1) < 5) or (substr(key, 1, 1) >= 5)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((substr(key, 1, 1) < 5) or (substr(key, 1, 1) >= 5)) (type: boolean) + Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: substr(key, 1, 1) (type: string), key (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: substr(key, 1, 1) (type: string) + Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 332 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (KEY._col0 >= 5) (type: boolean) + Statistics: Num rows: 110 Data size: 19360 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g2_n4 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, c1 + Statistics: Num rows: 105 Data size: 9345 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: (KEY._col0 < 5) (type: boolean) + Statistics: Num rows: 110 Data size: 19360 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 105 Data size: 10605 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g3_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) + outputColumnNames: key, c1, c2 + Statistics: Num rows: 105 Data size: 9765 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll'), compute_stats(c2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g2_n4 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, c1 + Column Types: string, int + Table: default.dest_g2_n4 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_g3_n0 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, c1, c2 + Column Types: string, int, int + Table: default.dest_g3_n0 + +PREHOOK: query: FROM src +INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) +INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest_g2_n4 +PREHOOK: Output: default@dest_g3_n0 +POSTHOOK: query: FROM src +INSERT OVERWRITE TABLE dest_g2_n4 SELECT substr(src.key,1,1), count(DISTINCT src.key) WHERE substr(src.key,1,1) >= 5 GROUP BY substr(src.key,1,1) +INSERT OVERWRITE TABLE dest_g3_n0 SELECT substr(src.key,1,1), count(DISTINCT src.key), count(src.value) WHERE substr(src.key,1,1) < 5 GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest_g2_n4 +POSTHOOK: Output: default@dest_g3_n0 +POSTHOOK: Lineage: dest_g2_n4.c1 EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: dest_g2_n4.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_g3_n0.c1 EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: dest_g3_n0.c2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_g3_n0.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT * FROM dest_g2_n4 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_g2_n4 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM dest_g2_n4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_g2_n4 +#### A masked pattern was here #### +5 6 +6 5 +7 6 +8 8 +9 7 +PREHOOK: query: SELECT * FROM dest_g3_n0 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_g3_n0 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM dest_g3_n0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_g3_n0 +#### A masked pattern was here #### +0 1 3 +1 71 115 +2 69 111 +3 62 99 +4 74 124 +PREHOOK: query: DROP TABLE dest_g2_n4 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@dest_g2_n4 +PREHOOK: Output: default@dest_g2_n4 +POSTHOOK: query: DROP TABLE dest_g2_n4 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@dest_g2_n4 +POSTHOOK: Output: default@dest_g2_n4 +PREHOOK: query: DROP TABLE dest_g3_n0 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@dest_g3_n0 +PREHOOK: Output: default@dest_g3_n0 +POSTHOOK: query: DROP TABLE dest_g3_n0 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@dest_g3_n0 +POSTHOOK: Output: default@dest_g3_n0 diff --git ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer3.q.out ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer3.q.out new file mode 100644 index 0000000000..4d4e87c3ad --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_multi_single_reducer3.q.out @@ -0,0 +1,1054 @@ +PREHOOK: query: create table e1_n1 (key string, count int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@e1_n1 +POSTHOOK: query: create table e1_n1 (key string, count int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@e1_n1 +PREHOOK: query: create table e2_n2 (key string, count int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: create table e2_n2 (key string, count int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@e2_n2 +PREHOOK: query: explain +from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: explain +from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((VALUE._col0) IN ('val_100', 'val_200', 'val_300') and (KEY._col0) IN (100, 150, 200)) (type: boolean) + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: ((VALUE._col0) IN ('val_400', 'val_500') and (KEY._col0) IN (400, 450)) (type: boolean) + Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e1_n1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e2_n2 + +PREHOOK: query: from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from e1_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from e1_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +100 2 +200 2 +PREHOOK: query: select * from e2_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +POSTHOOK: query: select * from e2_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +400 1 +PREHOOK: query: explain +from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: explain +from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: value (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: value (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((VALUE._col0 + VALUE._col0) = 200) or ((VALUE._col0 - 100) = 100) or ((VALUE._col0 = 300) and KEY._col0 is not null)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: (((VALUE._col0 + VALUE._col0) = 400) or (((VALUE._col0 - 100) = 500) and KEY._col0 is not null)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e1_n1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e2_n2 + +PREHOOK: query: from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from e1_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from e1_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +val_100 2 +val_200 2 +PREHOOK: query: select * from e2_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +POSTHOOK: query: select * from e2_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +val_200 2 +PREHOOK: query: explain +from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: explain +from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((value) IN ('val_400', 'val_500') and (key) IN (400, 450)) or ((value) IN ('val_100', 'val_200', 'val_300') and (key) IN (100, 150, 200))) (type: boolean) + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((VALUE._col0) IN ('val_100', 'val_200', 'val_300') and (KEY._col0) IN (100, 150, 200)) (type: boolean) + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 285 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 3 Data size: 273 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: ((VALUE._col0) IN ('val_400', 'val_500') and (KEY._col0) IN (400, 450)) (type: boolean) + Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 2 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e1_n1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e2_n2 + +PREHOOK: query: from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: from src +insert overwrite table e1_n1 +select key, count(*) +where src.value in ('val_100', 'val_200', 'val_300') AND key in (100, 150, 200) +group by key +insert overwrite table e2_n2 +select key, count(*) +where src.value in ('val_400', 'val_500') AND key in (400, 450) +group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from e1_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from e1_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +100 2 +200 2 +PREHOOK: query: select * from e2_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +POSTHOOK: query: select * from e2_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +400 1 +PREHOOK: query: explain +from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: explain +from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((((key + key) = 400) or (((key - 100) = 500) and value is not null)) or (((key + key) = 200) or ((key - 100) = 100) or ((key = 300) and value is not null))) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: value (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: value (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((VALUE._col0 + VALUE._col0) = 200) or ((VALUE._col0 - 100) = 100) or ((VALUE._col0 = 300) and KEY._col0 is not null)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: (((VALUE._col0 + VALUE._col0) = 400) or (((VALUE._col0 - 100) = 500) and KEY._col0 is not null)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, count + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(count, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e1_n1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e1_n1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.e2_n2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, count + Column Types: string, int + Table: default.e2_n2 + +PREHOOK: query: from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@e1_n1 +PREHOOK: Output: default@e2_n2 +POSTHOOK: query: from src +insert overwrite table e1_n1 +select value, count(*) +where src.key + src.key = 200 or src.key - 100 = 100 or src.key = 300 AND VALUE IS NOT NULL +group by value +insert overwrite table e2_n2 +select value, count(*) +where src.key + src.key = 400 or src.key - 100 = 500 AND VALUE IS NOT NULL +group by value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@e1_n1 +POSTHOOK: Output: default@e2_n2 +POSTHOOK: Lineage: e1_n1.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e1_n1.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: e2_n2.count EXPRESSION [(src)src.null, ] +POSTHOOK: Lineage: e2_n2.key SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from e1_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from e1_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e1_n1 +#### A masked pattern was here #### +val_100 2 +val_200 2 +PREHOOK: query: select * from e2_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +POSTHOOK: query: select * from e2_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@e2_n2 +#### A masked pattern was here #### +val_200 2 diff --git ql/src/test/results/clientpositive/llap/groupby_multialias.q.out ql/src/test/results/clientpositive/llap/groupby_multialias.q.out new file mode 100644 index 0000000000..0c4966ac6b --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_multialias.q.out @@ -0,0 +1,82 @@ +PREHOOK: query: create table t1_n150 (a int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t1_n150 +POSTHOOK: query: create table t1_n150 (a int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t1_n150 +PREHOOK: query: explain +select t1_n150.a as a1, min(t1_n150.a) as a +from t1_n150 +group by t1_n150.a +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n150 +#### A masked pattern was here #### +POSTHOOK: query: explain +select t1_n150.a as a1, min(t1_n150.a) as a +from t1_n150 +group by t1_n150.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n150 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n150 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: int) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: min(a) + keys: a (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: min(VALUE._col0) + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/groupby_nocolumnalign.q.out ql/src/test/results/clientpositive/llap/groupby_nocolumnalign.q.out similarity index 60% rename from ql/src/test/results/clientpositive/groupby_nocolumnalign.q.out rename to ql/src/test/results/clientpositive/llap/groupby_nocolumnalign.q.out index 3a92e71a75..77c888fd92 100644 --- ql/src/test/results/clientpositive/groupby_nocolumnalign.q.out +++ ql/src/test/results/clientpositive/llap/groupby_nocolumnalign.q.out @@ -20,70 +20,101 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@dest2_n31_2 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), substr(value, 5) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), substr(value, 5) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 132000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 500 Data size: 132000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT _col2) - keys: _col0 (type: string), _col1 (type: string), _col2 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 500 Data size: 136000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col2:0._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n31_2 - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, val1, val2 - Statistics: Num rows: 500 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 132000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col2) + keys: _col0 (type: string), _col1 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n31_2 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 250 Data size: 69750 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -95,7 +126,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n31_2 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -103,30 +134,6 @@ STAGE PLANS: Column Types: int, string, string Table: default.dest2_n31_2 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM SRC INSERT OVERWRITE TABLE DEST2_n31_2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) GROUP BY SRC.key, SRC.value PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/groupby_position.q.out ql/src/test/results/clientpositive/llap/groupby_position.q.out new file mode 100644 index 0000000000..1d6ed69bed --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_position.q.out @@ -0,0 +1,1850 @@ +PREHOOK: query: CREATE TABLE testTable1(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@testTable1 +POSTHOOK: query: CREATE TABLE testTable1(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@testTable1 +PREHOOK: query: CREATE TABLE testTable2(key INT, val1 STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@testTable2 +POSTHOOK: query: CREATE TABLE testTable2(key INT, val1 STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@testTable2 +PREHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@testtable1 +PREHOOK: Output: default@testtable2 +POSTHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@testtable1 +POSTHOOK: Output: default@testtable2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 20) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 20) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), value (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 45152 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 166 Data size: 45152 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable1 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9879518 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 30876 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable2 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.testtable1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.testtable2 + +PREHOOK: query: FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@testtable1 +PREHOOK: Output: default@testtable2 +POSTHOOK: query: FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@testtable1 +POSTHOOK: Output: default@testtable2 +POSTHOOK: Lineage: testtable1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: testtable1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@testtable1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testtable1 +#### A masked pattern was here #### +0 1 +2 1 +4 1 +5 1 +8 1 +9 1 +10 1 +11 1 +12 1 +15 1 +17 1 +18 1 +19 1 +PREHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@testtable2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testtable2 +#### A masked pattern was here #### +0 val_0 1 +2 val_2 1 +4 val_4 1 +5 val_5 1 +8 val_8 1 +9 val_9 1 +10 val_10 1 +11 val_11 1 +12 val_12 1 +15 val_15 1 +17 val_17 1 +18 val_18 1 +19 val_19 1 +PREHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@testtable1 +PREHOOK: Output: default@testtable2 +POSTHOOK: query: EXPLAIN +FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@testtable1 +POSTHOOK: Output: default@testtable2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 20) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 83 Data size: 15023 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 20) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT substr(value, 5)) + keys: value (type: string), key (type: string), substr(value, 5) (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 45152 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 166 Data size: 45152 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable1 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9879518 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 30876 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col1) (type: int), _col0 (type: string), CAST( _col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable2 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, val1, val2 + Statistics: Num rows: 166 Data size: 46314 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable1 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.testtable1 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.testtable2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val1, val2 + Column Types: int, string, string + Table: default.testtable2 + +PREHOOK: query: FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@testtable1 +PREHOOK: Output: default@testtable2 +POSTHOOK: query: FROM SRC +INSERT OVERWRITE TABLE testTable1 SELECT SRC.key, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 1 +INSERT OVERWRITE TABLE testTable2 SELECT SRC.key, SRC.value, COUNT(DISTINCT SUBSTR(SRC.value,5)) WHERE SRC.key < 20 GROUP BY 2, 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@testtable1 +POSTHOOK: Output: default@testtable2 +POSTHOOK: Lineage: testtable1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: testtable1.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.val1 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: testtable2.val2 EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@testtable1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, value FROM testTable1 ORDER BY 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testtable1 +#### A masked pattern was here #### +0 1 +2 1 +4 1 +5 1 +8 1 +9 1 +10 1 +11 1 +12 1 +15 1 +17 1 +18 1 +19 1 +PREHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@testtable2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val1, val2 FROM testTable2 ORDER BY 1, 2, 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testtable2 +#### A masked pattern was here #### +0 val_0 1 +2 val_2 1 +4 val_4 1 +5 val_5 1 +8 val_8 1 +9 val_9 1 +10 val_10 1 +11 val_11 1 +12 val_12 1 +15 val_15 1 +17 val_17 1 +18 val_18 1 +19 val_19 1 +PREHOOK: query: EXPLAIN +SELECT t.key, t.value +FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t +ORDER BY 2 DESC, 1 ASC +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT t.key, t.value +FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t +ORDER BY 2 DESC, 1 ASC +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: b + filterExpr: (UDFToDouble(key) <= 20.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) <= 20.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: bigint), _col0 (type: string) + null sort order: zz + sort order: -+ + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 83 Data size: 7885 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT t.key, t.value +FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t +ORDER BY 2 DESC, 1 ASC +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT t.key, t.value +FROM (SELECT b.key as key, count(1) as value FROM src b WHERE b.key <= 20 GROUP BY 1) t +ORDER BY 2 DESC, 1 ASC +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 3 +5 3 +12 2 +15 2 +18 2 +10 1 +11 1 +17 1 +19 1 +2 1 +20 1 +4 1 +8 1 +9 1 +PREHOOK: query: EXPLAIN +SELECT c1, c2, c3, c4 +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 + ) a + JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT c1, c2, c3, c4 +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 + ) a + JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) + Reducer 5 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), substr(value, 5) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 7128 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 27 Data size: 7128 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), value (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 7128 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 7128 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 13 Data size: 2314 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 13 Data size: 2314 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 13 Data size: 4628 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: --++ + Statistics: Num rows: 13 Data size: 4628 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 13 Data size: 4628 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 13 Data size: 4628 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT c1, c2, c3, c4 +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 + ) a + JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT c1, c2, c3, c4 +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2, COUNT(DISTINCT SUBSTR(src1.value,5)) AS c3 WHERE src1.key > 10 and src1.key < 20 GROUP BY 1, 2 + ) a + JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 GROUP BY 1, 2 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +ORDER BY 1 DESC, 2 DESC, 3 ASC, 4 ASC +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +19 val_19 19 val_19 +18 val_18 18 val_18 +17 val_17 17 val_17 +PREHOOK: query: EXPLAIN +SELECT key FROM src ORDER BY 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key FROM src ORDER BY 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key FROM src ORDER BY 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT key FROM src ORDER BY 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 +0 +0 +10 +100 +100 +103 +103 +104 +104 +105 +11 +111 +113 +113 +114 +116 +118 +118 +119 +119 +119 +12 +12 +120 +120 +125 +125 +126 +128 +128 +128 +129 +129 +131 +133 +134 +134 +136 +137 +137 +138 +138 +138 +138 +143 +145 +146 +146 +149 +149 +15 +15 +150 +152 +152 +153 +155 +156 +157 +158 +160 +162 +163 +164 +164 +165 +165 +166 +167 +167 +167 +168 +169 +169 +169 +169 +17 +170 +172 +172 +174 +174 +175 +175 +176 +176 +177 +178 +179 +179 +18 +18 +180 +181 +183 +186 +187 +187 +187 +189 +19 +190 +191 +191 +192 +193 +193 +193 +194 +195 +195 +196 +197 +197 +199 +199 +199 +2 +20 +200 +200 +201 +202 +203 +203 +205 +205 +207 +207 +208 +208 +208 +209 +209 +213 +213 +214 +216 +216 +217 +217 +218 +219 +219 +221 +221 +222 +223 +223 +224 +224 +226 +228 +229 +229 +230 +230 +230 +230 +230 +233 +233 +235 +237 +237 +238 +238 +239 +239 +24 +24 +241 +242 +242 +244 +247 +248 +249 +252 +255 +255 +256 +256 +257 +258 +26 +26 +260 +262 +263 +265 +265 +266 +27 +272 +272 +273 +273 +273 +274 +275 +277 +277 +277 +277 +278 +278 +28 +280 +280 +281 +281 +282 +282 +283 +284 +285 +286 +287 +288 +288 +289 +291 +292 +296 +298 +298 +298 +30 +302 +305 +306 +307 +307 +308 +309 +309 +310 +311 +311 +311 +315 +316 +316 +316 +317 +317 +318 +318 +318 +321 +321 +322 +322 +323 +325 +325 +327 +327 +327 +33 +331 +331 +332 +333 +333 +335 +336 +338 +339 +34 +341 +342 +342 +344 +344 +345 +348 +348 +348 +348 +348 +35 +35 +35 +351 +353 +353 +356 +360 +362 +364 +365 +366 +367 +367 +368 +369 +369 +369 +37 +37 +373 +374 +375 +377 +378 +379 +382 +382 +384 +384 +384 +386 +389 +392 +393 +394 +395 +395 +396 +396 +396 +397 +397 +399 +399 +4 +400 +401 +401 +401 +401 +401 +402 +403 +403 +403 +404 +404 +406 +406 +406 +406 +407 +409 +409 +409 +41 +411 +413 +413 +414 +414 +417 +417 +417 +418 +419 +42 +42 +421 +424 +424 +427 +429 +429 +43 +430 +430 +430 +431 +431 +431 +432 +435 +436 +437 +438 +438 +438 +439 +439 +44 +443 +444 +446 +448 +449 +452 +453 +454 +454 +454 +455 +457 +458 +458 +459 +459 +460 +462 +462 +463 +463 +466 +466 +466 +467 +468 +468 +468 +468 +469 +469 +469 +469 +469 +47 +470 +472 +475 +477 +478 +478 +479 +480 +480 +480 +481 +482 +483 +484 +485 +487 +489 +489 +489 +489 +490 +491 +492 +492 +493 +494 +495 +496 +497 +498 +498 +498 +5 +5 +5 +51 +51 +53 +54 +57 +58 +58 +64 +65 +66 +67 +67 +69 +70 +70 +70 +72 +72 +74 +76 +76 +77 +78 +8 +80 +82 +83 +83 +84 +84 +85 +86 +87 +9 +90 +90 +90 +92 +95 +95 +96 +97 +97 +98 +98 +PREHOOK: query: EXPLAIN +SELECT distinct key FROM src ORDER BY 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT distinct key FROM src ORDER BY 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT distinct key FROM src ORDER BY 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT distinct key FROM src ORDER BY 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 +10 +100 +103 +104 +105 +11 +111 +113 +114 +116 +118 +119 +12 +120 +125 +126 +128 +129 +131 +133 +134 +136 +137 +138 +143 +145 +146 +149 +15 +150 +152 +153 +155 +156 +157 +158 +160 +162 +163 +164 +165 +166 +167 +168 +169 +17 +170 +172 +174 +175 +176 +177 +178 +179 +18 +180 +181 +183 +186 +187 +189 +19 +190 +191 +192 +193 +194 +195 +196 +197 +199 +2 +20 +200 +201 +202 +203 +205 +207 +208 +209 +213 +214 +216 +217 +218 +219 +221 +222 +223 +224 +226 +228 +229 +230 +233 +235 +237 +238 +239 +24 +241 +242 +244 +247 +248 +249 +252 +255 +256 +257 +258 +26 +260 +262 +263 +265 +266 +27 +272 +273 +274 +275 +277 +278 +28 +280 +281 +282 +283 +284 +285 +286 +287 +288 +289 +291 +292 +296 +298 +30 +302 +305 +306 +307 +308 +309 +310 +311 +315 +316 +317 +318 +321 +322 +323 +325 +327 +33 +331 +332 +333 +335 +336 +338 +339 +34 +341 +342 +344 +345 +348 +35 +351 +353 +356 +360 +362 +364 +365 +366 +367 +368 +369 +37 +373 +374 +375 +377 +378 +379 +382 +384 +386 +389 +392 +393 +394 +395 +396 +397 +399 +4 +400 +401 +402 +403 +404 +406 +407 +409 +41 +411 +413 +414 +417 +418 +419 +42 +421 +424 +427 +429 +43 +430 +431 +432 +435 +436 +437 +438 +439 +44 +443 +444 +446 +448 +449 +452 +453 +454 +455 +457 +458 +459 +460 +462 +463 +466 +467 +468 +469 +47 +470 +472 +475 +477 +478 +479 +480 +481 +482 +483 +484 +485 +487 +489 +490 +491 +492 +493 +494 +495 +496 +497 +498 +5 +51 +53 +54 +57 +58 +64 +65 +66 +67 +69 +70 +72 +74 +76 +77 +78 +8 +80 +82 +83 +84 +85 +86 +87 +9 +90 +92 +95 +96 +97 +98 diff --git ql/src/test/results/clientpositive/llap/groupby_ppd.q.out ql/src/test/results/clientpositive/llap/groupby_ppd.q.out new file mode 100644 index 0000000000..51c5c6abba --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_ppd.q.out @@ -0,0 +1,120 @@ +PREHOOK: query: create table invites_n0 (id int, foo int, bar int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@invites_n0 +POSTHOOK: query: create table invites_n0 (id int, foo int, bar int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@invites_n0 +PREHOOK: query: explain select * from (select foo, bar from (select bar, foo from invites_n0 c union all select bar, foo from invites_n0 d) b) a group by bar, foo having bar=1 +PREHOOK: type: QUERY +PREHOOK: Input: default@invites_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from (select foo, bar from (select bar, foo from invites_n0 c union all select bar, foo from invites_n0 d) b) a group by bar, foo having bar=1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@invites_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: c + filterExpr: (bar = 1) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (bar = 1) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: foo (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: d + filterExpr: (bar = 1) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (bar = 1) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: foo (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), 1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 2 + Vertex: Union 2 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: drop table invites_n0 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@invites_n0 +PREHOOK: Output: default@invites_n0 +POSTHOOK: query: drop table invites_n0 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@invites_n0 +POSTHOOK: Output: default@invites_n0 diff --git ql/src/test/results/clientpositive/llap/groupby_ppr.q.out ql/src/test/results/clientpositive/llap/groupby_ppr.q.out new file mode 100644 index 0000000000..b54f27c8bc --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_ppr.q.out @@ -0,0 +1,354 @@ +PREHOOK: query: CREATE TABLE dest1_n79(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n79 +POSTHOOK: query: CREATE TABLE dest1_n79(key STRING, c1 INT, c2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n79 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n79 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n79 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1_n79 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n79 +OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +GROUP BY SUBSTR(`key`, 1, 1) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [src] + /srcpart/ds=2008-04-08/hr=12 [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 31916 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2 + columns.comments + columns.types string:int:string +#### A masked pattern was here #### + name default.dest1_n79 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n79 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string) + outputColumnNames: key, c1, c2 + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 316 Data size: 86268 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: key (type: string), c1 (type: int), c2 (type: string) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll') + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2 + columns.comments + columns.types string:int:string +#### A masked pattern was here #### + name default.dest1_n79 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n79 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, c1, c2 + Column Types: string, int, string + Table: default.dest1_n79 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n79 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n79 +POSTHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1_n79 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n79 +POSTHOOK: Lineage: dest1_n79.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n79.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n79.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n79.* FROM dest1_n79 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n79 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n79.* FROM dest1_n79 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n79 +#### A masked pattern was here #### +0 1 00.0 +1 71 132828.0 +2 69 251142.0 +3 62 364008.0 +4 74 4105526.0 +5 6 5794.0 +6 5 6796.0 +7 6 71470.0 +8 8 81524.0 +9 7 92094.0 diff --git ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out new file mode 100644 index 0000000000..c307d9c25a --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out @@ -0,0 +1,704 @@ +PREHOOK: query: CREATE TABLE dest1(key STRING, c1 INT, c2 STRING, c3 INT, c4 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1 +POSTHOOK: query: CREATE TABLE dest1(key STRING, c1 INT, c2 STRING, c3 INT, c4 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1 +OPTIMIZED SQL: SELECT SUBSTR(`key`, 1, 1) AS `_o__c0`, COUNT(DISTINCT SUBSTR(`value`, 5)) AS `_o__c1`, SUBSTR(`key`, 1, 1) || SUM(SUBSTR(`value`, 5)) AS `_o__c2`, SUM(DISTINCT SUBSTR(`value`, 5)) AS `_o__c3`, COUNT(DISTINCT `value`) AS `_o__c4` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +GROUP BY SUBSTR(`key`, 1, 1) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + null sort order: zzz + numBuckets: -1 + sort order: +++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [src] + /srcpart/ds=2008-04-08/hr=12 [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 36972 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), concat(_col0, _col2) (type: string), UDFToInteger(_col3) (type: int), UDFToInteger(_col4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) + outputColumnNames: key, c1, c2, c3, c4 + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 316 Data size: 88796 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: key (type: string), c1 (type: int), c2 (type: string), c3 (type: int), c4 (type: int) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll'), compute_stats(VALUE._col4, 'hll'), compute_stats(VALUE._col5, 'hll') + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types struct:struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, c1, c2, c3, c4 + Column Types: string, int, string, int, int + Table: default.dest1 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1 +POSTHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1 +POSTHOOK: Lineage: dest1.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.c3 EXPRESSION [(srcpart)src.null, ] +POSTHOOK: Lineage: dest1.c4 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1.* FROM dest1 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1.* FROM dest1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1 +#### A masked pattern was here #### +0 1 00.0 0 1 +1 71 132828.0 10044 71 +2 69 251142.0 15780 69 +3 62 364008.0 20119 62 +4 74 4105526.0 30965 74 +5 6 5794.0 278 6 +6 5 6796.0 331 5 +7 6 71470.0 447 6 +8 8 81524.0 595 8 +9 7 92094.0 577 7 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: substr(key, 1, 1) (type: string), substr(value, 5) (type: string), value (type: string) + outputColumnNames: $f0, $f1, $f2 + Statistics: Num rows: 1000 Data size: 459000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: $f0 (type: string), $f1 (type: string), $f2 (type: string) + null sort order: zzz + numBuckets: -1 + sort order: +++ + Map-reduce partition columns: $f0 (type: string) + Statistics: Num rows: 1000 Data size: 459000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [src] + /srcpart/ds=2008-04-08/hr=12 [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), sum(KEY._col1:0._col0), sum(DISTINCT KEY._col1:1._col0), count(DISTINCT KEY._col1:2._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: $f0, $f1, $f2, $f3, $f4 + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: $f0 (type: string), UDFToInteger($f1) (type: int), concat($f0, $f2) (type: string), UDFToInteger($f3) (type: int), UDFToInteger($f4) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1 + numFiles 1 + numRows 10 + rawDataSize 184 + serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 194 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) + outputColumnNames: key, c1, c2, c3, c4 + Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 196 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: key (type: string), c1 (type: int), c2 (type: string), c3 (type: int), c4 (type: int) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0, 'hll'), compute_stats(VALUE._col2, 'hll'), compute_stats(VALUE._col3, 'hll'), compute_stats(VALUE._col4, 'hll'), compute_stats(VALUE._col5, 'hll') + mode: complete + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types struct:struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,c1,c2,c3,c4 + columns.comments + columns.types string:int:string:int:int +#### A masked pattern was here #### + name default.dest1 + numFiles 1 + numRows 10 + rawDataSize 184 + serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 194 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, c1, c2, c3, c4 + Column Types: string, int, string, int, int + Table: default.dest1 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1 +POSTHOOK: query: FROM srcpart src +INSERT OVERWRITE TABLE dest1 +SELECT substr(src.key,1,1), count(DISTINCT substr(src.value,5)), concat(substr(src.key,1,1),sum(substr(src.value,5))), sum(DISTINCT substr(src.value, 5)), count(DISTINCT src.value) +WHERE src.ds = '2008-04-08' +GROUP BY substr(src.key,1,1) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1 +POSTHOOK: Lineage: dest1.c1 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.c2 EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), (srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.c3 EXPRESSION [(srcpart)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1.c4 EXPRESSION [(srcpart)src.null, ] +POSTHOOK: Lineage: dest1.key EXPRESSION [(srcpart)src.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1.* FROM dest1 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1.* FROM dest1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1 +#### A masked pattern was here #### +0 1 00.0 0 1 +1 71 132828.0 10044 71 +2 69 251142.0 15780 69 +3 62 364008.0 20119 62 +4 74 4105526.0 30965 74 +5 6 5794.0 278 6 +6 5 6796.0 331 5 +7 6 71470.0 447 6 +8 8 81524.0 595 8 +9 7 92094.0 577 7 diff --git ql/src/test/results/clientpositive/llap/groupby_rollup1.q.out ql/src/test/results/clientpositive/llap/groupby_rollup1.q.out new file mode 100644 index 0000000000..675f58eeeb --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_rollup1.q.out @@ -0,0 +1,708 @@ +PREHOOK: query: CREATE TABLE T1_n91(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n91 +POSTHOOK: query: CREATE TABLE T1_n91(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n91 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n91 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n91 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n91 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n91 +PREHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n91 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +1 11 1 +1 NULL 1 +2 12 1 +2 NULL 1 +3 13 1 +3 NULL 1 +7 17 1 +7 NULL 1 +8 18 1 +8 28 1 +8 NULL 2 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n91 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(DISTINCT val) + keys: key (type: string), 0L (type: bigint), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: bigint) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col2 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n91 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, val, count(1) FROM T1_n91 GROUP BY key, val with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +1 11 1 +1 NULL 1 +2 12 1 +2 NULL 1 +3 13 1 +3 NULL 1 +7 17 1 +7 NULL 1 +8 18 1 +8 28 1 +8 NULL 2 +NULL NULL 6 +PREHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n91 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(DISTINCT val) + keys: key (type: string), 0L (type: bigint), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint), _col2 (type: string) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col2:0._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: bigint) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: bigint) + mode: final + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col2 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +POSTHOOK: query: SELECT key, count(distinct val) FROM T1_n91 GROUP BY key with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +NULL 6 +PREHOOK: query: CREATE TABLE T2_n56(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n56 +POSTHOOK: query: CREATE TABLE T2_n56(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n56 +PREHOOK: query: CREATE TABLE T3_n20(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n20 +POSTHOOK: query: CREATE TABLE T3_n20(key1 STRING, key2 STRING, val INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n20 +PREHOOK: query: EXPLAIN +FROM T1_n91 +INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup +INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by rollup(key, val) +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +PREHOOK: Output: default@t2_n56 +PREHOOK: Output: default@t3_n20 +POSTHOOK: query: EXPLAIN +FROM T1_n91 +INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup +INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by rollup(key, val) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +POSTHOOK: Output: default@t2_n56 +POSTHOOK: Output: default@t3_n20 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 1 (SIMPLE_EDGE) + Reducer 6 <- Reducer 5 (SIMPLE_EDGE) + Reducer 7 <- Reducer 6 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n91 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(1) + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(1) + keys: key (type: string), val (type: string), 0L (type: bigint) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n56 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: partials + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col3 (type: bigint) + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: final + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col3) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t3_n20 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, val + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(val, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1680 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 7 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1696 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n56 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key1, key2, val + Column Types: string, string, int + Table: default.t2_n56 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t3_n20 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key1, key2, val + Column Types: string, string, int + Table: default.t3_n20 + +PREHOOK: query: FROM T1_n91 +INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup +INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by key, val with rollup +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n91 +PREHOOK: Output: default@t2_n56 +PREHOOK: Output: default@t3_n20 +POSTHOOK: query: FROM T1_n91 +INSERT OVERWRITE TABLE T2_n56 SELECT key, val, count(1) group by key, val with rollup +INSERT OVERWRITE TABLE T3_n20 SELECT key, val, sum(1) group by key, val with rollup +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n91 +POSTHOOK: Output: default@t2_n56 +POSTHOOK: Output: default@t3_n20 +POSTHOOK: Lineage: t2_n56.key1 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n56.key2 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n56.val EXPRESSION [(t1_n91)t1_n91.null, ] +POSTHOOK: Lineage: t3_n20.key1 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t3_n20.key2 SIMPLE [(t1_n91)t1_n91.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: t3_n20.val EXPRESSION [(t1_n91)t1_n91.null, ] diff --git ql/src/test/results/clientpositive/groupby_sort_10.q.out ql/src/test/results/clientpositive/llap/groupby_sort_10.q.out similarity index 66% rename from ql/src/test/results/clientpositive/groupby_sort_10.q.out rename to ql/src/test/results/clientpositive/llap/groupby_sort_10.q.out index 570d3eeeaf..d72a3a715e 100644 --- ql/src/test/results/clientpositive/groupby_sort_10.q.out +++ ql/src/test/results/clientpositive/llap/groupby_sort_10.q.out @@ -31,39 +31,23 @@ POSTHOOK: Input: default@t1_n149 POSTHOOK: Input: default@t1_n149@ds=1 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n149 - Statistics: Num rows: 4 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 4 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: string) - mode: final - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: t1_n149 + Select Operator + expressions: key (type: string) + outputColumnNames: key + Group By Operator + keys: key (type: string) + mode: final + outputColumnNames: _col0 + ListSink PREHOOK: query: select distinct key from T1_n149 PREHOOK: type: QUERY @@ -107,42 +91,51 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n149 - Statistics: Num rows: 8 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 8 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n149 + Statistics: Num rows: 8 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 8 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.75 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/groupby_sort_11.q.out ql/src/test/results/clientpositive/llap/groupby_sort_11.q.out new file mode 100644 index 0000000000..eca8131457 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_11.q.out @@ -0,0 +1,644 @@ +PREHOOK: query: CREATE TABLE T1_n18(key STRING, val STRING) PARTITIONED BY (ds string) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n18 +POSTHOOK: query: CREATE TABLE T1_n18(key STRING, val STRING) PARTITIONED BY (ds string) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n18 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n18 PARTITION (ds='1') +SELECT * from src where key < 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@t1_n18@ds=1 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n18 PARTITION (ds='1') +SELECT * from src where key < 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@t1_n18@ds=1 +POSTHOOK: Lineage: t1_n18 PARTITION(ds=1).key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: t1_n18 PARTITION(ds=1).val SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: EXPLAIN select count(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string) + mode: final + outputColumnNames: _col0 + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col0) + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +6 +PREHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: double) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0), count(VALUE._col1), count(VALUE._col2), sum(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +6 10 10 28.0 +PREHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint), _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col1), count(VALUE._col2), sum(DISTINCT KEY._col1:1._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: bigint), _col2 (type: bigint), _col3 (type: bigint), _col4 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 5 Data size: 160 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5 Data size: 160 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +1 1 1 2.0 +1 1 1 4.0 +1 1 1 8.0 +1 1 1 9.0 +1 3 3 0.0 +1 3 3 5.0 +PREHOOK: query: EXPLAIN select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(DISTINCT key), count(), count(key), sum(DISTINCT key) + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint), _col3 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(DISTINCT KEY._col1:0._col0), count(VALUE._col1), count(VALUE._col2), sum(DISTINCT KEY._col1:1._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5 Data size: 585 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select key, count(distinct key), count(1), count(key), sum(distinct key) from T1_n18 group by key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +0 1 3 3 0.0 +2 1 1 1 2.0 +4 1 1 1 4.0 +5 1 3 3 5.0 +8 1 1 1 8.0 +9 1 1 1 9.0 +PREHOOK: query: EXPLAIN select count(distinct key+key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct key+key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: double) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 5 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col0) + minReductionHashAggr: 0.8 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct key+key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct key+key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +6 +PREHOOK: query: EXPLAIN select count(distinct 1) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct 1) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 10 Data size: 150 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: true (type: boolean) + minReductionHashAggr: 0.9 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: boolean) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: boolean) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct 1) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct 1) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +1 +PREHOOK: query: EXPLAIN select count(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN select count(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n18 + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 10 Data size: 850 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0 + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: complete + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(distinct key) from T1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n18 +PREHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +POSTHOOK: query: select count(distinct key) from T1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n18 +POSTHOOK: Input: default@t1_n18@ds=1 +#### A masked pattern was here #### +6 diff --git ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out new file mode 100644 index 0000000000..023929da92 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out @@ -0,0 +1,5704 @@ +PREHOOK: query: CREATE TABLE T1_n80(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n80 +POSTHOOK: query: CREATE TABLE T1_n80(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n80 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n80 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n80 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n80 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n80 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n80 select key, val from T1_n80 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@t1_n80 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n80 select key, val from T1_n80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@t1_n80 +POSTHOOK: Lineage: t1_n80.key SIMPLE [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n80.val SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n18(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n18 +POSTHOOK: query: CREATE TABLE outputTbl1_n18(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n18 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T1_n80 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T1_n80 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T1_n80 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T1_n80 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: CREATE TABLE outputTbl2_n5(key1 int, key2 string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl2_n5 +POSTHOOK: query: CREATE TABLE outputTbl2_n5(key1 int, key2 string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl2_n5 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl2_n5 +SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl2_n5 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl2_n5 +SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl2_n5 +OPTIMIZED SQL: SELECT `key`, `val`, COUNT(*) AS `$f2` +FROM `default`.`t1_n80` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:string:int +#### A masked pattern was here #### + name default.outputtbl2_n5 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n5 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:string:int +#### A masked pattern was here #### + name default.outputtbl2_n5 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n5 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, string, int + Table: default.outputtbl2_n5 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n5 +SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl2_n5 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n5 +SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl2_n5 +POSTHOOK: Lineage: outputtbl2_n5.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl2_n5.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl2_n5.key2 SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl2_n5 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl2_n5 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl2_n5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl2_n5 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n80) subq1 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n80) subq1 GROUP BY k +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: CREATE TABLE outputTbl3_n2(key1 int, key2 int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl3_n2 +POSTHOOK: query: CREATE TABLE outputTbl3_n2(key1 int, key2 int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl3_n2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl3_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl3_n2 +OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, COUNT(*) AS `_o__c2` +FROM `default`.`t1_n80` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: 1 (type: int), UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, int, int + Table: default.outputtbl3_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl3_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT 1, key, count(1) FROM T1_n80 GROUP BY 1, key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl3_n2 +POSTHOOK: Lineage: outputtbl3_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl3_n2.key1 SIMPLE [] +POSTHOOK: Lineage: outputtbl3_n2.key2 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl3_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl3_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl3_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl3_n2 +#### A masked pattern was here #### +1 1 1 +1 2 1 +1 3 1 +1 7 1 +1 8 2 +PREHOOK: query: CREATE TABLE outputTbl4_n2(key1 int, key2 int, key3 string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl4_n2 +POSTHOOK: query: CREATE TABLE outputTbl4_n2(key1 int, key2 int, key3 string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl4_n2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl4_n2 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t1_n80` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T1_n80 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl4_n2 +POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl3_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl3_n2 +OPTIMIZED SQL: SELECT `key` AS `$f0`, CAST(`key` AS DOUBLE) + 1 AS `$f1`, COUNT(*) AS `$f2` +FROM `default`.`t1_n80` +GROUP BY `key`, CAST(`key` AS DOUBLE) + 1 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: double) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n2 + numFiles 1 + numRows 5 + rawDataSize 25 + serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n2 + numFiles 1 + numRows 5 + rawDataSize 25 + serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, int, int + Table: default.outputtbl3_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl3_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n2 +SELECT key, key + 1, count(1) FROM T1_n80 GROUP BY key, key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl3_n2 +POSTHOOK: Lineage: outputtbl3_n2.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl3_n2.key1 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl3_n2.key2 EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl3_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl3_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl3_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl3_n2 +#### A masked pattern was here #### +1 2 1 +2 3 1 +3 4 1 +7 8 1 +8 9 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key + key, sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +group by key + key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key + key, sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +group by key + key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS `$f0`, SUM(COUNT(*)) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key + key, sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +group by key + key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key + key, sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +group by key + key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +14 1 +16 2 +2 1 +4 1 +6 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n80 GROUP BY key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n80 GROUP BY key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +UNION ALL +SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Map 4 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n80 GROUP BY key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n80 GROUP BY key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +1 1 +2 1 +2 1 +3 1 +3 1 +7 1 +7 1 +8 2 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) FROM T1_n80 GROUP BY key + key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n80 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) FROM T1_n80 GROUP BY key + key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +GROUP BY `key` +UNION ALL +SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `key`, COUNT(*) AS `_o__c1` +FROM `default`.`t1_n80` +GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 4 (SIMPLE_EDGE), Union 2 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Map 4 + Map Operator Tree: + TableScan + alias: t1_n80 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 5 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: double) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: CAST( _col0 AS STRING) (type: string), _col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n80 GROUP BY key + key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT * FROM ( +SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n80 GROUP BY key + key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +14 1 +16 2 +2 1 +2 1 +3 1 +4 1 +6 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`$f1` + `t2`.`$f1` AS `_o__c1` +FROM (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t0` +INNER JOIN (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Dummy Store + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Map Operator Tree: + TableScan + alias: t1_n80 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger((_col1 + _col3)) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 3 + numRows 10 + rawDataSize 32 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 42 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 3 + numRows 10 + rawDataSize 32 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 42 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n80 GROUP BY key) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t1_n80)t1_n80.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 2 +2 2 +3 2 +7 2 +PREHOOK: query: EXPLAIN EXTENDED +SELECT * FROM +(SELECT key, count(1) FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED +SELECT * FROM +(SELECT key, count(1) FROM T1_n80 GROUP BY key) subq1 +JOIN +(SELECT key, val, count(1) FROM T1_n80 GROUP BY key, val) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n80` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t0` +INNER JOIN (SELECT `key`, `val`, COUNT(*) AS `$f2` +FROM `default`.`t1_n80` +WHERE `key` IS NOT NULL +GROUP BY `key`, `val`) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n80 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + tag: 0 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Map 3 + Map Operator Tree: + TableScan + alias: t1_n80 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n80 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n80 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n80 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n80 + name: default.t1_n80 + Truncated Path -> Alias: + /t1_n80 [t1_n80] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Position of Big Table: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types string:bigint:string:string:bigint + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 4 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: 1 + value expressions: _col1 (type: string), _col2 (type: bigint) + auto parallelism: true + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: CREATE TABLE T2_n49(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n49 +POSTHOOK: query: CREATE TABLE T2_n49(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n49 +PREHOOK: query: INSERT OVERWRITE TABLE T2_n49 select key, val from T1_n80 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n80 +PREHOOK: Output: default@t2_n49 +POSTHOOK: query: INSERT OVERWRITE TABLE T2_n49 select key, val from T1_n80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n80 +POSTHOOK: Output: default@t2_n49 +POSTHOOK: Lineage: t2_n49.key SIMPLE [(t1_n80)t1_n80.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n49.val SIMPLE [(t1_n80)t1_n80.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T2_n49 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T2_n49 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl1_n18 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t2_n49` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n49 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n49 + name: default.t2_n49 + Truncated Path -> Alias: + /t2_n49 [t2_n49] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 4 + rawDataSize 12 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 16 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n18 + numFiles 1 + numRows 4 + rawDataSize 12 + serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 16 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n18 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T2_n49 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n18 +SELECT key, count(1) FROM T2_n49 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl1_n18 +POSTHOOK: Lineage: outputtbl1_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: outputtbl1_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n18 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n49` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n49 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n49 + name: default.t2_n49 + Truncated Path -> Alias: + /t2_n49 [t2_n49] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, 1, val, count(1) FROM T2_n49 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: CREATE TABLE outputTbl5_n2(key1 int, key2 int, key3 string, key4 int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl5_n2 +POSTHOOK: query: CREATE TABLE outputTbl5_n2(key1 int, key2 int, key3 string, key4 int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl5_n2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl5_n2 +SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl5_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl5_n2 +SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl5_n2 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, 2 AS `_o__c3`, COUNT(*) AS `_o__c4` +FROM `default`.`t2_n49` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), 2 (type: int), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,key4,cnt + columns.comments + columns.types int:int:string:int:int +#### A masked pattern was here #### + name default.outputtbl5_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl5_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) + outputColumnNames: key1, key2, key3, key4, cnt + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(key4, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n49 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n49 + name: default.t2_n49 + Truncated Path -> Alias: + /t2_n49 [t2_n49] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types struct:struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,key4,cnt + columns.comments + columns.types int:int:string:int:int +#### A masked pattern was here #### + name default.outputtbl5_n2 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl5_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, key4, cnt + Column Types: int, int, string, int, int + Table: default.outputtbl5_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n2 +SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl5_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n2 +SELECT key, 1, val, 2, count(1) FROM T2_n49 GROUP BY key, 1, val, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl5_n2 +POSTHOOK: Lineage: outputtbl5_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: outputtbl5_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl5_n2.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl5_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl5_n2.key4 SIMPLE [] +PREHOOK: query: SELECT * FROM outputTbl5_n2 +ORDER BY key1, key2, key3, key4 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl5_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl5_n2 +ORDER BY key1, key2, key3, key4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl5_n2 +#### A masked pattern was here #### +1 1 11 2 1 +2 1 12 2 1 +3 1 13 2 1 +7 1 17 2 1 +8 1 18 2 1 +8 1 28 2 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n49)subq +group by key, constant, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n49)subq +group by key, constant, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +OPTIMIZED SQL: SELECT `key`, 1 AS `constant`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n49` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n49 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n49 + name: default.t2_n49 + Truncated Path -> Alias: + /t2_n49 [t2_n49] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n49)subq +group by key, constant, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n49)subq +group by key, constant, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n49)subq +)subq2 +group by key, constant3, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n2 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n49)subq +)subq2 +group by key, constant3, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +OPTIMIZED SQL: SELECT `key`, 2 AS `constant3`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n49` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 2 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n49 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n49 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n49 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n49 + name: default.t2_n49 + Truncated Path -> Alias: + /t2_n49 [t2_n49] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n2 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n2 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n49)subq +)subq2 +group by key, constant3, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@outputtbl4_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n2 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n49)subq +)subq2 +group by key, constant3, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@outputtbl4_n2 +POSTHOOK: Lineage: outputtbl4_n2.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: outputtbl4_n2.key1 EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n2.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n2.key3 SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n2 +#### A masked pattern was here #### +1 2 11 1 +2 2 12 1 +3 2 13 1 +7 2 17 1 +8 2 18 1 +8 2 28 1 +PREHOOK: query: CREATE TABLE DEST1_n80(key INT, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST1_n80 +POSTHOOK: query: CREATE TABLE DEST1_n80(key INT, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST1_n80 +PREHOOK: query: CREATE TABLE DEST2_n18(key INT, val STRING, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST2_n18 +POSTHOOK: query: CREATE TABLE DEST2_n18(key INT, val STRING, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST2_n18 +PREHOOK: query: EXPLAIN +FROM T2_n49 +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@dest1_n80 +PREHOOK: Output: default@dest2_n18 +POSTHOOK: query: EXPLAIN +FROM T2_n49 +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@dest1_n80 +POSTHOOK: Output: default@dest2_n18 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n18 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n80 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n80 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest1_n80 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n18 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: int, string, int + Table: default.dest2_n18 + +PREHOOK: query: FROM T2_n49 +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@dest1_n80 +PREHOOK: Output: default@dest2_n18 +POSTHOOK: query: FROM T2_n49 +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@dest1_n80 +POSTHOOK: Output: default@dest2_n18 +POSTHOOK: Lineage: dest1_n80.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: dest1_n80.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: dest2_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n18.val SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: select * from DEST1_n80 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n80 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST1_n80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n80 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: select * from DEST2_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n18 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST2_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n18 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 +PREHOOK: query: EXPLAIN +FROM (select key, val from T2_n49 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@dest1_n80 +PREHOOK: Output: default@dest2_n18 +POSTHOOK: query: EXPLAIN +FROM (select key, val from T2_n49 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@dest1_n80 +POSTHOOK: Output: default@dest2_n18 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n49 + filterExpr: (UDFToDouble(key) = 8.0D) (type: boolean) + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) = 8.0D) (type: boolean) + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + bucketGroup: true + keys: _col0 (type: string) + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Group By Operator + aggregations: count(1) + keys: _col0 (type: string), _col1 (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 537 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n18 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n80 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n80 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest1_n80 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n18 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: int, string, int + Table: default.dest2_n18 + +PREHOOK: query: FROM (select key, val from T2_n49 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n49 +PREHOOK: Output: default@dest1_n80 +PREHOOK: Output: default@dest2_n18 +POSTHOOK: query: FROM (select key, val from T2_n49 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n80 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n18 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n49 +POSTHOOK: Output: default@dest1_n80 +POSTHOOK: Output: default@dest2_n18 +POSTHOOK: Lineage: dest1_n80.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: dest1_n80.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n18.cnt EXPRESSION [(t2_n49)t2_n49.null, ] +POSTHOOK: Lineage: dest2_n18.key EXPRESSION [(t2_n49)t2_n49.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n18.val SIMPLE [(t2_n49)t2_n49.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: select * from DEST1_n80 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n80 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST1_n80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n80 +#### A masked pattern was here #### +8 2 +PREHOOK: query: select * from DEST2_n18 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n18 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST2_n18 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n18 +#### A masked pattern was here #### +8 18 1 +8 28 1 diff --git ql/src/test/results/clientpositive/llap/groupby_sort_2.q.out ql/src/test/results/clientpositive/llap/groupby_sort_2.q.out new file mode 100644 index 0000000000..7b01cc8e1d --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_2.q.out @@ -0,0 +1,188 @@ +PREHOOK: query: CREATE TABLE T1_n51(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (val) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n51 +POSTHOOK: query: CREATE TABLE T1_n51(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (val) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n51 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n51 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n51 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n51 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n51 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n51 select key, val from T1_n51 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n51 +PREHOOK: Output: default@t1_n51 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n51 select key, val from T1_n51 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n51 +POSTHOOK: Output: default@t1_n51 +POSTHOOK: Lineage: t1_n51.key SIMPLE [(t1_n51)t1_n51.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n51.val SIMPLE [(t1_n51)t1_n51.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n10(val string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n10 +POSTHOOK: query: CREATE TABLE outputTbl1_n10(val string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n10 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n10 +SELECT val, count(1) FROM T1_n51 GROUP BY val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n51 +PREHOOK: Output: default@outputtbl1_n10 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n10 +SELECT val, count(1) FROM T1_n51 GROUP BY val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n51 +POSTHOOK: Output: default@outputtbl1_n10 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n51 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val (type: string) + outputColumnNames: val + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + bucketGroup: true + keys: val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n10 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: val, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n10 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: val, cnt + Column Types: string, int + Table: default.outputtbl1_n10 + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n10 +SELECT val, count(1) FROM T1_n51 GROUP BY val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n51 +PREHOOK: Output: default@outputtbl1_n10 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n10 +SELECT val, count(1) FROM T1_n51 GROUP BY val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n51 +POSTHOOK: Output: default@outputtbl1_n10 +POSTHOOK: Lineage: outputtbl1_n10.cnt EXPRESSION [(t1_n51)t1_n51.null, ] +POSTHOOK: Lineage: outputtbl1_n10.val SIMPLE [(t1_n51)t1_n51.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n10 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n10 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n10 +#### A masked pattern was here #### +11 1 +12 1 +13 1 +17 1 +18 1 +28 1 diff --git ql/src/test/results/clientpositive/llap/groupby_sort_3.q.out ql/src/test/results/clientpositive/llap/groupby_sort_3.q.out new file mode 100644 index 0000000000..0c34c3dc5b --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_3.q.out @@ -0,0 +1,310 @@ +PREHOOK: query: CREATE TABLE T1_n89(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n89 +POSTHOOK: query: CREATE TABLE T1_n89(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n89 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n89 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n89 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n89 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n89 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n89 select key, val from T1_n89 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n89 +PREHOOK: Output: default@t1_n89 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n89 select key, val from T1_n89 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n89 +POSTHOOK: Output: default@t1_n89 +POSTHOOK: Lineage: t1_n89.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n89.val SIMPLE [(t1_n89)t1_n89.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n20(key string, val string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n20 +POSTHOOK: query: CREATE TABLE outputTbl1_n20(key string, val string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n20 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n20 +SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n89 +PREHOOK: Output: default@outputtbl1_n20 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n20 +SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n89 +POSTHOOK: Output: default@outputtbl1_n20 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n89 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n20 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n20 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: string, string, int + Table: default.outputtbl1_n20 + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n20 +SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n89 +PREHOOK: Output: default@outputtbl1_n20 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n20 +SELECT key, val, count(1) FROM T1_n89 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n89 +POSTHOOK: Output: default@outputtbl1_n20 +POSTHOOK: Lineage: outputtbl1_n20.cnt EXPRESSION [(t1_n89)t1_n89.null, ] +POSTHOOK: Lineage: outputtbl1_n20.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl1_n20.val SIMPLE [(t1_n89)t1_n89.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n20 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n20 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n20 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 +PREHOOK: query: CREATE TABLE outputTbl2_n7(key string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl2_n7 +POSTHOOK: query: CREATE TABLE outputTbl2_n7(key string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl2_n7 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl2_n7 +SELECT key, count(1) FROM T1_n89 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n89 +PREHOOK: Output: default@outputtbl2_n7 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl2_n7 +SELECT key, count(1) FROM T1_n89 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n89 +POSTHOOK: Output: default@outputtbl2_n7 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n89 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n7 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n7 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: string, int + Table: default.outputtbl2_n7 + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n7 +SELECT key, count(1) FROM T1_n89 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n89 +PREHOOK: Output: default@outputtbl2_n7 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n7 +SELECT key, count(1) FROM T1_n89 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n89 +POSTHOOK: Output: default@outputtbl2_n7 +POSTHOOK: Lineage: outputtbl2_n7.cnt EXPRESSION [(t1_n89)t1_n89.null, ] +POSTHOOK: Lineage: outputtbl2_n7.key SIMPLE [(t1_n89)t1_n89.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl2_n7 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl2_n7 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl2_n7 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl2_n7 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 diff --git ql/src/test/results/clientpositive/llap/groupby_sort_4.q.out ql/src/test/results/clientpositive/llap/groupby_sort_4.q.out new file mode 100644 index 0000000000..656b2547ad --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_4.q.out @@ -0,0 +1,347 @@ +PREHOOK: query: CREATE TABLE T1_n133(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n133 +POSTHOOK: query: CREATE TABLE T1_n133(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n133 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n133 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n133 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n133 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n133 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n133 select key, val from T1_n133 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n133 +PREHOOK: Output: default@t1_n133 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n133 select key, val from T1_n133 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n133 +POSTHOOK: Output: default@t1_n133 +POSTHOOK: Lineage: t1_n133.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n133.val SIMPLE [(t1_n133)t1_n133.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n31(key STRING, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n31 +POSTHOOK: query: CREATE TABLE outputTbl1_n31(key STRING, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n31 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n31 +SELECT key, count(1) FROM T1_n133 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n133 +PREHOOK: Output: default@outputtbl1_n31 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n31 +SELECT key, count(1) FROM T1_n133 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n133 +POSTHOOK: Output: default@outputtbl1_n31 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n133 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n31 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n31 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: string, int + Table: default.outputtbl1_n31 + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n31 +SELECT key, count(1) FROM T1_n133 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n133 +PREHOOK: Output: default@outputtbl1_n31 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n31 +SELECT key, count(1) FROM T1_n133 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n133 +POSTHOOK: Output: default@outputtbl1_n31 +POSTHOOK: Lineage: outputtbl1_n31.cnt EXPRESSION [(t1_n133)t1_n133.null, ] +POSTHOOK: Lineage: outputtbl1_n31.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n31 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n31 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n31 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n31 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: CREATE TABLE outputTbl2_n8(key STRING, val STRING, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl2_n8 +POSTHOOK: query: CREATE TABLE outputTbl2_n8(key STRING, val STRING, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl2_n8 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl2_n8 +SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n133 +PREHOOK: Output: default@outputtbl2_n8 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl2_n8 +SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n133 +POSTHOOK: Output: default@outputtbl2_n8 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n133 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n8 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n8 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: string, string, int + Table: default.outputtbl2_n8 + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n8 +SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n133 +PREHOOK: Output: default@outputtbl2_n8 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n8 +SELECT key, val, count(1) FROM T1_n133 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n133 +POSTHOOK: Output: default@outputtbl2_n8 +POSTHOOK: Lineage: outputtbl2_n8.cnt EXPRESSION [(t1_n133)t1_n133.null, ] +POSTHOOK: Lineage: outputtbl2_n8.key SIMPLE [(t1_n133)t1_n133.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl2_n8.val SIMPLE [(t1_n133)t1_n133.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl2_n8 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl2_n8 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl2_n8 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl2_n8 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 diff --git ql/src/test/results/clientpositive/groupby_sort_5.q.out ql/src/test/results/clientpositive/llap/groupby_sort_5.q.out similarity index 54% rename from ql/src/test/results/clientpositive/groupby_sort_5.q.out rename to ql/src/test/results/clientpositive/llap/groupby_sort_5.q.out index 90312062f9..cbb258d66e 100644 --- ql/src/test/results/clientpositive/groupby_sort_5.q.out +++ ql/src/test/results/clientpositive/llap/groupby_sort_5.q.out @@ -48,80 +48,80 @@ POSTHOOK: Input: default@t1_n6 POSTHOOK: Output: default@outputtbl1_n5 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n6 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n6 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: count() + keys: key (type: string), val (type: string) + mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -133,7 +133,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n5 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -141,36 +141,6 @@ STAGE PLANS: Column Types: string, string, int Table: default.outputtbl1_n5 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n5 SELECT key, val, count(1) FROM T1_n6 GROUP BY key, val PREHOOK: type: QUERY @@ -248,80 +218,80 @@ POSTHOOK: Input: default@t1_n6 POSTHOOK: Output: default@outputtbl1_n5 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n6 - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n6 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: count() + keys: key (type: string), val (type: string) + mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -333,7 +303,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n5 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -341,36 +311,6 @@ STAGE PLANS: Column Types: string, string, int Table: default.outputtbl1_n5 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n5 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n5 SELECT key, val, count(1) FROM T1_n6 GROUP BY key, val PREHOOK: type: QUERY @@ -456,72 +396,99 @@ POSTHOOK: Input: default@t1_n6 POSTHOOK: Output: default@outputtbl2_n1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n6 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n6 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl2_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: int) - outputColumnNames: key, cnt - Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -533,7 +500,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2_n1 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -541,30 +508,6 @@ STAGE PLANS: Column Types: string, int Table: default.outputtbl2_n1 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1048 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n1 SELECT key, count(1) FROM T1_n6 GROUP BY key PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out new file mode 100644 index 0000000000..68f196cd03 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out @@ -0,0 +1,769 @@ +PREHOOK: query: CREATE TABLE T1_n61(key STRING, val STRING) PARTITIONED BY (ds string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n61 +POSTHOOK: query: CREATE TABLE T1_n61(key STRING, val STRING) PARTITIONED BY (ds string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n61 +PREHOOK: query: CREATE TABLE outputTbl1_n15(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n15 +POSTHOOK: query: CREATE TABLE outputTbl1_n15(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n15 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Output: default@outputtbl1_n15 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n61` +WHERE `ds` = '1' +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n61 + filterExpr: (ds = '1') (type: boolean) + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (ds = '1') (type: boolean) + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: unknown + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n15 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Output: default@outputtbl1_n15 +POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] +POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n15 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n15 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n61 PARTITION (ds='2') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n61 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n61 PARTITION (ds='2') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n61 +POSTHOOK: Output: default@t1_n61@ds=2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Output: default@outputtbl1_n15 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n61` +WHERE `ds` = '1' +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n61 + filterExpr: (ds = '1') (type: boolean) + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (ds = '1') (type: boolean) + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: unknown + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 92 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n15 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '1' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Output: default@outputtbl1_n15 +POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] +POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n15 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n15 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Input: default@t1_n61@ds=2 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Input: default@t1_n61@ds=2 +POSTHOOK: Output: default@outputtbl1_n15 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n61` +WHERE `ds` = '2' +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n61 + filterExpr: (ds = '2') (type: boolean) + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: ds=2 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2 + properties: + bucket_count -1 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n61 + numFiles 1 + numRows 0 + partition_columns ds + partition_columns.types string + rawDataSize 0 + serialization.ddl struct t1_n61 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n61 + partition_columns ds + partition_columns.types string + serialization.ddl struct t1_n61 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n61 + name: default.t1_n61 + Truncated Path -> Alias: + /t1_n61/ds=2 [t1_n61] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 184 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: PARTIAL Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: PARTIAL Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: PARTIAL Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n15 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n15 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n15 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n61 +PREHOOK: Input: default@t1_n61@ds=2 +PREHOOK: Output: default@outputtbl1_n15 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n15 +SELECT key, count(1) FROM T1_n61 where ds = '2' GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n61 +POSTHOOK: Input: default@t1_n61@ds=2 +POSTHOOK: Output: default@outputtbl1_n15 +POSTHOOK: Lineage: outputtbl1_n15.cnt EXPRESSION [(t1_n61)t1_n61.null, ] +POSTHOOK: Lineage: outputtbl1_n15.key EXPRESSION [(t1_n61)t1_n61.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n15 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n15 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n15 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 diff --git ql/src/test/results/clientpositive/groupby_sort_7.q.out ql/src/test/results/clientpositive/llap/groupby_sort_7.q.out similarity index 53% rename from ql/src/test/results/clientpositive/groupby_sort_7.q.out rename to ql/src/test/results/clientpositive/llap/groupby_sort_7.q.out index a0a193d720..5e785ffeb5 100644 --- ql/src/test/results/clientpositive/groupby_sort_7.q.out +++ ql/src/test/results/clientpositive/llap/groupby_sort_7.q.out @@ -53,81 +53,81 @@ POSTHOOK: Input: default@t1_n104@ds=1 POSTHOOK: Output: default@outputtbl1_n26 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n104 - filterExpr: (ds = '1') (type: boolean) - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string), val (type: string) - outputColumnNames: key, val - Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: key (type: string), val (type: string) - mode: final - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n26 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n104 + filterExpr: (ds = '1') (type: boolean) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) - outputColumnNames: key, val, cnt - Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: count() + keys: key (type: string), val (type: string) + mode: final outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n26 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1672 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -139,7 +139,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n26 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -147,36 +147,6 @@ STAGE PLANS: Column Types: string, string, int Table: default.outputtbl1_n26 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n26 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1_n26 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n26 SELECT key, val, count(1) FROM T1_n104 where ds = '1' GROUP BY key, val PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/groupby_sort_8.q.out ql/src/test/results/clientpositive/llap/groupby_sort_8.q.out similarity index 57% rename from ql/src/test/results/clientpositive/groupby_sort_8.q.out rename to ql/src/test/results/clientpositive/llap/groupby_sort_8.q.out index b5f581e6e6..4eef22a03f 100644 --- ql/src/test/results/clientpositive/groupby_sort_8.q.out +++ ql/src/test/results/clientpositive/llap/groupby_sort_8.q.out @@ -47,42 +47,54 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n45 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count(DISTINCT key) - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n45 Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: key (type: string) + mode: final + outputColumnNames: _col0 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(_col0) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/groupby_sort_9.q.out ql/src/test/results/clientpositive/llap/groupby_sort_9.q.out similarity index 64% rename from ql/src/test/results/clientpositive/groupby_sort_9.q.out rename to ql/src/test/results/clientpositive/llap/groupby_sort_9.q.out index 33e21a3e08..dc4d40db46 100644 --- ql/src/test/results/clientpositive/groupby_sort_9.q.out +++ ql/src/test/results/clientpositive/llap/groupby_sort_9.q.out @@ -61,45 +61,54 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t1_n96 - Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n96 + Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - aggregations: count() - bucketGroup: true - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 2208 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out new file mode 100644 index 0000000000..5b9ee467a0 --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out @@ -0,0 +1,5892 @@ +PREHOOK: query: CREATE TABLE T1_n56(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n56 +POSTHOOK: query: CREATE TABLE T1_n56(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n56 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n56 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n56 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n56 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n56 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n56 select key, val from T1_n56 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@t1_n56 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n56 select key, val from T1_n56 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@t1_n56 +POSTHOOK: Lineage: t1_n56.key SIMPLE [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n56.val SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n13(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n13 +POSTHOOK: query: CREATE TABLE outputTbl1_n13(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n13 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T1_n56 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T1_n56 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T1_n56 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T1_n56 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: CREATE TABLE outputTbl2_n3(key1 int, key2 string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl2_n3 +POSTHOOK: query: CREATE TABLE outputTbl2_n3(key1 int, key2 string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl2_n3 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl2_n3 +SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl2_n3 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl2_n3 +SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl2_n3 +OPTIMIZED SQL: SELECT `key`, `val`, COUNT(*) AS `$f2` +FROM `default`.`t1_n56` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:string:int +#### A masked pattern was here #### + name default.outputtbl2_n3 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n3 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1656 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:string:int +#### A masked pattern was here #### + name default.outputtbl2_n3 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl2_n3 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, string, int + Table: default.outputtbl2_n3 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n3 +SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl2_n3 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl2_n3 +SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl2_n3 +POSTHOOK: Lineage: outputtbl2_n3.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl2_n3.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl2_n3.key2 SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl2_n3 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl2_n3 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl2_n3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl2_n3 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM (SELECT key, val FROM T1_n56) subq1 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT k, count(1) FROM (SELECT key as k, val as v FROM T1_n56) subq1 GROUP BY k +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: CREATE TABLE outputTbl3_n1(key1 int, key2 int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl3_n1 +POSTHOOK: query: CREATE TABLE outputTbl3_n1(key1 int, key2 int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl3_n1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl3_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl3_n1 +OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, COUNT(*) AS `_o__c2` +FROM `default`.`t1_n56` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: 1 (type: int), UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, int, int + Table: default.outputtbl3_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl3_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT 1, key, count(1) FROM T1_n56 GROUP BY 1, key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl3_n1 +POSTHOOK: Lineage: outputtbl3_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl3_n1.key1 SIMPLE [] +POSTHOOK: Lineage: outputtbl3_n1.key2 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl3_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl3_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl3_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl3_n1 +#### A masked pattern was here #### +1 1 1 +1 2 1 +1 3 1 +1 7 1 +1 8 2 +PREHOOK: query: CREATE TABLE outputTbl4_n1(key1 int, key2 int, key3 string, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl4_n1 +POSTHOOK: query: CREATE TABLE outputTbl4_n1(key1 int, key2 int, key3 string, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl4_n1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl4_n1 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t1_n56` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 2080 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: final + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2128 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T1_n56 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl4_n1 +POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl3_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl3_n1 +OPTIMIZED SQL: SELECT `key` AS `$f0`, CAST(`key` AS DOUBLE) + 1 AS `$f1`, COUNT(*) AS `$f2` +FROM `default`.`t1_n56` +GROUP BY `key`, CAST(`key` AS DOUBLE) + 1 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: string), _col1 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: double) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: double) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: double) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: double) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n1 + numFiles 1 + numRows 5 + rawDataSize 25 + serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: key1, key2, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1456 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1504 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,cnt + columns.comments + columns.types int:int:int +#### A masked pattern was here #### + name default.outputtbl3_n1 + numFiles 1 + numRows 5 + rawDataSize 25 + serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, cnt + Column Types: int, int, int + Table: default.outputtbl3_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl3_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl3_n1 +SELECT key, key + 1, count(1) FROM T1_n56 GROUP BY key, key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl3_n1 +POSTHOOK: Lineage: outputtbl3_n1.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl3_n1.key1 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl3_n1.key2 EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl3_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl3_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl3_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl3_n1 +#### A masked pattern was here #### +1 2 1 +2 3 1 +3 4 1 +7 8 1 +8 9 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT cast(key + key as string), sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +group by key + key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT cast(key + key as string), sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +group by key + key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `_o__c0`, SUM(COUNT(*)) AS `_o__c1` +FROM `default`.`t1_n56` +GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: (UDFToDouble(_col0) + UDFToDouble(_col0)) (type: double), _col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: double) + mode: partials + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: double) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(CAST( _col0 AS STRING)) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 5 + rawDataSize 15 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 20 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT cast(key + key as string), sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +group by key + key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT cast(key + key as string), sum(cnt) from +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +group by key + key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +14 1 +16 2 +2 1 +4 1 +6 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n56 GROUP BY key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n56 GROUP BY key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +UNION ALL +SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Map 4 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 5 + rawDataSize 17 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 22 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n56 GROUP BY key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT key, count(1) FROM T1_n56 GROUP BY key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +1 1 +2 1 +2 1 +3 1 +3 1 +7 1 +7 1 +8 2 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) FROM T1_n56 GROUP BY key + key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) FROM T1_n56 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) FROM T1_n56 GROUP BY key + key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +GROUP BY `key` +UNION ALL +SELECT CAST(CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) AS STRING) AS `key`, COUNT(*) AS `_o__c1` +FROM `default`.`t1_n56` +GROUP BY CAST(`key` AS DOUBLE) + CAST(`key` AS DOUBLE) +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 4 (SIMPLE_EDGE) + Reducer 6 <- Reducer 5 (SIMPLE_EDGE), Union 2 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Map 4 + Map Operator Tree: + TableScan + alias: t1_n56 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Select Operator + expressions: (UDFToDouble(key) + UDFToDouble(key)) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: double) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 5 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: double) + mode: partials + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: double) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Reducer 6 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: double) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: CAST( _col0 AS STRING) (type: string), _col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 2 + numRows 10 + rawDataSize 30 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 40 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n56 GROUP BY key + key +) subq1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT * FROM ( +SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key + UNION ALL +SELECT cast(key + key as string) as key, count(1) as cnt FROM T1_n56 GROUP BY key + key +) subq1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +14 1 +16 2 +2 1 +2 1 +3 1 +4 1 +6 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`$f1` + `t2`.`$f1` AS `_o__c1` +FROM (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t0` +INNER JOIN (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Dummy Store + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Map Operator Tree: + TableScan + alias: t1_n56 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger((_col1 + _col3)) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 3 + numRows 10 + rawDataSize 32 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 42 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 3 + numRows 10 + rawDataSize 32 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 42 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT subq1.key, subq1.cnt+subq2.cnt FROM +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, count(1) as cnt FROM T1_n56 GROUP BY key) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t1_n56)t1_n56.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 2 +2 2 +3 2 +7 2 +PREHOOK: query: EXPLAIN EXTENDED +SELECT * FROM +(SELECT key, count(1) FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val) subq2 +ON subq1.key = subq2.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED +SELECT * FROM +(SELECT key, count(1) FROM T1_n56 GROUP BY key) subq1 +JOIN +(SELECT key, val, count(1) FROM T1_n56 GROUP BY key, val) subq2 +ON subq1.key = subq2.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t1_n56` +WHERE `key` IS NOT NULL +GROUP BY `key`) AS `t0` +INNER JOIN (SELECT `key`, `val`, COUNT(*) AS `$f2` +FROM `default`.`t1_n56` +WHERE `key` IS NOT NULL +GROUP BY `key`, `val`) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n56 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + tag: 0 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Map 3 + Map Operator Tree: + TableScan + alias: t1_n56 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t1_n56 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t1_n56 + numFiles 1 + numRows 6 + rawDataSize 24 + serialization.ddl struct t1_n56 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t1_n56 + name: default.t1_n56 + Truncated Path -> Alias: + /t1_n56 [t1_n56] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Position of Big Table: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 607 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types string:bigint:string:string:bigint + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 4 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: partials + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 6 Data size: 2208 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col2 (type: bigint) + auto parallelism: true + Reducer 5 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + tag: 1 + value expressions: _col1 (type: string), _col2 (type: bigint) + auto parallelism: true + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: CREATE TABLE T2_n34(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n34 +POSTHOOK: query: CREATE TABLE T2_n34(key STRING, val STRING) +CLUSTERED BY (key, val) SORTED BY (key, val) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n34 +PREHOOK: query: INSERT OVERWRITE TABLE T2_n34 select key, val from T1_n56 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n56 +PREHOOK: Output: default@t2_n34 +POSTHOOK: query: INSERT OVERWRITE TABLE T2_n34 select key, val from T1_n56 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n56 +POSTHOOK: Output: default@t2_n34 +POSTHOOK: Lineage: t2_n34.key SIMPLE [(t1_n56)t1_n56.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n34.val SIMPLE [(t1_n56)t1_n56.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T2_n34 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T2_n34 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl1_n13 +OPTIMIZED SQL: SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`t2_n34` +GROUP BY `key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n34 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n34 + name: default.t2_n34 + Truncated Path -> Alias: + /t2_n34 [t2_n34] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: partials + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 4 + rawDataSize 12 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 16 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,cnt + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.outputtbl1_n13 + numFiles 1 + numRows 4 + rawDataSize 12 + serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 16 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n13 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T2_n34 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl1_n13 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl1_n13 +SELECT key, count(1) FROM T2_n34 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl1_n13 +POSTHOOK: Lineage: outputtbl1_n13.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: outputtbl1_n13.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl1_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl1_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl1_n13 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n34` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n34 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n34 + name: default.t2_n34 + Truncated Path -> Alias: + /t2_n34 [t2_n34] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: final + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, 1, val, count(1) FROM T2_n34 GROUP BY key, 1, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: CREATE TABLE outputTbl5_n1(key1 int, key2 int, key3 string, key4 int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl5_n1 +POSTHOOK: query: CREATE TABLE outputTbl5_n1(key1 int, key2 int, key3 string, key4 int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl5_n1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl5_n1 +SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl5_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl5_n1 +SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl5_n1 +OPTIMIZED SQL: SELECT `key`, 1 AS `_o__c1`, `val`, 2 AS `_o__c3`, COUNT(*) AS `_o__c4` +FROM `default`.`t2_n34` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), 2 (type: int), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,key4,cnt + columns.comments + columns.types int:int:string:int:int +#### A masked pattern was here #### + name default.outputtbl5_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl5_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int), _col4 (type: int) + outputColumnNames: key1, key2, key3, key4, cnt + Statistics: Num rows: 6 Data size: 612 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(key4, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 2136 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n34 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n34 + name: default.t2_n34 + Truncated Path -> Alias: + /t2_n34 [t2_n34] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) + mode: final + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2200 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4 + columns.types struct:struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,key4,cnt + columns.comments + columns.types int:int:string:int:int +#### A masked pattern was here #### + name default.outputtbl5_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl5_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, key4, cnt + Column Types: int, int, string, int, int + Table: default.outputtbl5_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n1 +SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl5_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl5_n1 +SELECT key, 1, val, 2, count(1) FROM T2_n34 GROUP BY key, 1, val, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl5_n1 +POSTHOOK: Lineage: outputtbl5_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: outputtbl5_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl5_n1.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl5_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl5_n1.key4 SIMPLE [] +PREHOOK: query: SELECT * FROM outputTbl5_n1 +ORDER BY key1, key2, key3, key4 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl5_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl5_n1 +ORDER BY key1, key2, key3, key4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl5_n1 +#### A masked pattern was here #### +1 1 11 2 1 +2 1 12 2 1 +3 1 13 2 1 +7 1 17 2 1 +8 1 18 2 1 +8 1 28 2 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n34)subq +group by key, constant, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n34)subq +group by key, constant, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +OPTIMIZED SQL: SELECT `key`, 1 AS `constant`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n34` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 1 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n34 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n34 + name: default.t2_n34 + Truncated Path -> Alias: + /t2_n34 [t2_n34] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: final + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n34)subq +group by key, constant, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +SELECT key, constant, val, count(1) from +(SELECT key, 1 as constant, val from T2_n34)subq +group by key, constant, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +1 1 11 1 +2 1 12 1 +3 1 13 1 +7 1 17 1 +8 1 18 1 +8 1 28 1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n34)subq +)subq2 +group by key, constant3, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE outputTbl4_n1 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n34)subq +)subq2 +group by key, constant3, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +OPTIMIZED SQL: SELECT `key`, 2 AS `constant3`, `val`, COUNT(*) AS `_o__c3` +FROM `default`.`t2_n34` +GROUP BY `key`, `val` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), 2 (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: string), _col3 (type: int) + outputColumnNames: key1, key2, key3, cnt + Statistics: Num rows: 6 Data size: 588 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(key2, 'hll'), compute_stats(key3, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: t2_n34 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} + SORTBUCKETCOLSPREFIX TRUE + bucket_count 2 + bucket_field_name key,val + bucketing_version 2 + column.name.delimiter , + columns key,val + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.t2_n34 + numFiles 2 + numRows 6 + rawDataSize 24 + serialization.ddl struct t2_n34 { string key, string val} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 30 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n34 + name: default.t2_n34 + Truncated Path -> Alias: + /t2_n34 [t2_n34] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: final + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,key2,key3,cnt + columns.comments + columns.types int:int:string:int +#### A masked pattern was here #### + name default.outputtbl4_n1 + numFiles 2 + numRows 6 + rawDataSize 48 + serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 54 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl4_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key1, key2, key3, cnt + Column Types: int, int, string, int + Table: default.outputtbl4_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n34)subq +)subq2 +group by key, constant3, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@outputtbl4_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE outputTbl4_n1 +select key, constant3, val, count(1) from +( +SELECT key, constant as constant2, val, 2 as constant3 from +(SELECT key, 1 as constant, val from T2_n34)subq +)subq2 +group by key, constant3, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@outputtbl4_n1 +POSTHOOK: Lineage: outputtbl4_n1.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: outputtbl4_n1.key1 EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: outputtbl4_n1.key2 SIMPLE [] +POSTHOOK: Lineage: outputtbl4_n1.key3 SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM outputTbl4_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM outputTbl4_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@outputtbl4_n1 +#### A masked pattern was here #### +1 2 11 1 +2 2 12 1 +3 2 13 1 +7 2 17 1 +8 2 18 1 +8 2 28 1 +PREHOOK: query: CREATE TABLE DEST1_n57(key INT, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST1_n57 +POSTHOOK: query: CREATE TABLE DEST1_n57(key INT, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST1_n57 +PREHOOK: query: CREATE TABLE DEST2_n12(key INT, val STRING, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@DEST2_n12 +POSTHOOK: query: CREATE TABLE DEST2_n12(key INT, val STRING, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@DEST2_n12 +PREHOOK: query: EXPLAIN +FROM T2_n34 +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@dest1_n57 +PREHOOK: Output: default@dest2_n12 +POSTHOOK: query: EXPLAIN +FROM T2_n34 +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@dest1_n57 +POSTHOOK: Output: default@dest2_n12 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + bucketGroup: true + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: key, val + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + keys: key (type: string), val (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 1074 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n12 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: partials + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 279 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n57 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n57 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest1_n57 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n12 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: int, string, int + Table: default.dest2_n12 + +PREHOOK: query: FROM T2_n34 +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@dest1_n57 +PREHOOK: Output: default@dest2_n12 +POSTHOOK: query: FROM T2_n34 +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@dest1_n57 +POSTHOOK: Output: default@dest2_n12 +POSTHOOK: Lineage: dest1_n57.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: dest1_n57.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n12.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: dest2_n12.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n12.val SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: select * from DEST1_n57 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n57 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST1_n57 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n57 +#### A masked pattern was here #### +1 1 +2 1 +3 1 +7 1 +8 2 +PREHOOK: query: select * from DEST2_n12 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n12 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST2_n12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n12 +#### A masked pattern was here #### +1 11 1 +2 12 1 +3 13 1 +7 17 1 +8 18 1 +8 28 1 +PREHOOK: query: EXPLAIN +FROM (select key, val from T2_n34 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@dest1_n57 +PREHOOK: Output: default@dest2_n12 +POSTHOOK: query: EXPLAIN +FROM (select key, val from T2_n34 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@dest1_n57 +POSTHOOK: Output: default@dest2_n12 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t2_n34 + filterExpr: (UDFToDouble(key) = 8.0D) (type: boolean) + Statistics: Num rows: 6 Data size: 1026 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) = 8.0D) (type: boolean) + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), val (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(1) + bucketGroup: true + keys: _col0 (type: string) + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: rand() (type: double) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Group By Operator + aggregations: count(1) + keys: _col0 (type: string), _col1 (type: string) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 537 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n12 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, val, cnt + Statistics: Num rows: 3 Data size: 282 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(val, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: partials + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n57 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: final + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: true + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n57 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest1_n57 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n12 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, val, cnt + Column Types: int, string, int + Table: default.dest2_n12 + +PREHOOK: query: FROM (select key, val from T2_n34 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n34 +PREHOOK: Output: default@dest1_n57 +PREHOOK: Output: default@dest2_n12 +POSTHOOK: query: FROM (select key, val from T2_n34 where key = 8) x +INSERT OVERWRITE TABLE DEST1_n57 SELECT key, count(1) GROUP BY key +INSERT OVERWRITE TABLE DEST2_n12 SELECT key, val, count(1) GROUP BY key, val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n34 +POSTHOOK: Output: default@dest1_n57 +POSTHOOK: Output: default@dest2_n12 +POSTHOOK: Lineage: dest1_n57.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: dest1_n57.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n12.cnt EXPRESSION [(t2_n34)t2_n34.null, ] +POSTHOOK: Lineage: dest2_n12.key EXPRESSION [(t2_n34)t2_n34.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n12.val SIMPLE [(t2_n34)t2_n34.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: select * from DEST1_n57 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n57 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST1_n57 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n57 +#### A masked pattern was here #### +8 2 +PREHOOK: query: select * from DEST2_n12 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n12 +#### A masked pattern was here #### +POSTHOOK: query: select * from DEST2_n12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n12 +#### A masked pattern was here #### +8 18 1 +8 28 1 diff --git ql/src/test/results/clientpositive/llap/groupby_sort_test_1.q.out ql/src/test/results/clientpositive/llap/groupby_sort_test_1.q.out new file mode 100644 index 0000000000..a1ad55a9fb --- /dev/null +++ ql/src/test/results/clientpositive/llap/groupby_sort_test_1.q.out @@ -0,0 +1,143 @@ +PREHOOK: query: CREATE TABLE T1_n164(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n164 +POSTHOOK: query: CREATE TABLE T1_n164(key STRING, val STRING) +CLUSTERED BY (key) SORTED BY (key) INTO 2 BUCKETS STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n164 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n164 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n164 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/bucket_files/000000_0' INTO TABLE T1_n164 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n164 +PREHOOK: query: INSERT OVERWRITE TABLE T1_n164 select key, val from T1_n164 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n164 +PREHOOK: Output: default@t1_n164 +POSTHOOK: query: INSERT OVERWRITE TABLE T1_n164 select key, val from T1_n164 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n164 +POSTHOOK: Output: default@t1_n164 +POSTHOOK: Lineage: t1_n164.key SIMPLE [(t1_n164)t1_n164.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t1_n164.val SIMPLE [(t1_n164)t1_n164.FieldSchema(name:val, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE outputTbl1_n35(key int, cnt int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@outputTbl1_n35 +POSTHOOK: query: CREATE TABLE outputTbl1_n35(key int, cnt int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@outputTbl1_n35 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n35 +SELECT key, count(1) FROM T1_n164 GROUP BY key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n164 +PREHOOK: Output: default@outputtbl1_n35 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE outputTbl1_n35 +SELECT key, count(1) FROM T1_n164 GROUP BY key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n164 +POSTHOOK: Output: default@outputtbl1_n35 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n164 + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 6 Data size: 1104 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n35 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 3 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1032 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1064 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1_n35 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.outputtbl1_n35 + diff --git ql/src/test/results/clientpositive/groupingset_high_columns.q.out ql/src/test/results/clientpositive/llap/groupingset_high_columns.q.out similarity index 100% rename from ql/src/test/results/clientpositive/groupingset_high_columns.q.out rename to ql/src/test/results/clientpositive/llap/groupingset_high_columns.q.out index a7c47cd453..e10370e5d4 100644 --- ql/src/test/results/clientpositive/groupingset_high_columns.q.out +++ ql/src/test/results/clientpositive/llap/groupingset_high_columns.q.out @@ -128,38 +128,38 @@ POSTHOOK: query: select * from groupingsets32 POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets32 #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 3 NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 PREHOOK: query: create table groupingsets32a as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -271,38 +271,38 @@ POSTHOOK: query: select * from groupingsets32a POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets32a #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 2147483647 2147483647 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3221225471 3221225471 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3758096383 3758096383 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4026531839 4026531839 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4160749567 4160749567 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4227858431 4227858431 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4261412863 4261412863 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4278190079 4278190079 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4286578687 4286578687 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4290772991 4290772991 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4292870143 4292870143 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4293918719 4293918719 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294443007 4294443007 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294705151 4294705151 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294836223 4294836223 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294901759 4294901759 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294934527 4294934527 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294950911 4294950911 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294959103 4294959103 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294963199 4294963199 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294965247 4294965247 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966271 4294966271 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966783 4294966783 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967039 4294967039 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 4294967167 4294967167 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 4294967231 4294967231 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL 1 4294967263 4294967263 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL 1 4294967279 4294967279 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL 1 4294967287 4294967287 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 3 NULL NULL 1 4294967291 4294967291 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL 1 4294967293 4294967293 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 1 4294967294 4294967294 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 4294967231 4294967231 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 4294967167 4294967167 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967039 4294967039 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966783 4294966783 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966271 4294966271 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294965247 4294965247 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294963199 4294963199 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294959103 4294959103 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294950911 4294950911 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294934527 4294934527 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294901759 4294901759 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294836223 4294836223 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294705151 4294705151 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294443007 4294443007 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4293918719 4293918719 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4292870143 4292870143 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4290772991 4290772991 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4286578687 4286578687 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4278190079 4278190079 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4261412863 4261412863 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4227858431 4227858431 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4160749567 4160749567 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4026531839 4026531839 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3758096383 3758096383 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3221225471 3221225471 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 2147483647 2147483647 PREHOOK: query: create table groupingsets33 as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -401,32 +401,6 @@ POSTHOOK: query: select * from groupingsets33 POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets33 #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL NULL 1 @@ -434,6 +408,32 @@ NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 6 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 PREHOOK: query: create table groupingsets33a as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -546,32 +546,6 @@ POSTHOOK: query: select * from groupingsets33a POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets33a #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967295 4294967295 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 6442450943 6442450943 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 7516192767 7516192767 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8053063679 8053063679 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8321499135 8321499135 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8455716863 8455716863 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8522825727 8522825727 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8556380159 8556380159 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8573157375 8573157375 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8581545983 8581545983 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8585740287 8585740287 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8587837439 8587837439 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8588886015 8588886015 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589410303 8589410303 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589672447 8589672447 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589803519 8589803519 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589869055 8589869055 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589901823 8589901823 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589918207 8589918207 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589926399 8589926399 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589930495 8589930495 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589932543 8589932543 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589933567 8589933567 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934079 8589934079 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934335 8589934335 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 8589934463 8589934463 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL NULL 1 8589934527 8589934527 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL 1 8589934559 8589934559 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL NULL 1 8589934575 8589934575 @@ -579,6 +553,32 @@ NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL NULL 1 8589934587 8589934587 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 NULL 1 8589934589 8589934589 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 6 1 8589934590 8589934590 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 8589934463 8589934463 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934335 8589934335 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934079 8589934079 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589933567 8589933567 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589932543 8589932543 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589930495 8589930495 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589926399 8589926399 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589918207 8589918207 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589901823 8589901823 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589869055 8589869055 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589803519 8589803519 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589672447 8589672447 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589410303 8589410303 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8588886015 8588886015 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8587837439 8587837439 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8585740287 8585740287 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8581545983 8581545983 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8573157375 8573157375 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8556380159 8556380159 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8522825727 8522825727 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8455716863 8455716863 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8321499135 8321499135 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8053063679 8053063679 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 7516192767 7516192767 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 6442450943 6442450943 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967295 4294967295 PREHOOK: query: drop table groupingsets32 PREHOOK: type: DROPTABLE PREHOOK: Input: default@groupingsets32 @@ -716,38 +716,38 @@ POSTHOOK: query: select * from groupingsets32 POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets32 #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 3 NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 PREHOOK: query: create table groupingsets32a as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -859,38 +859,38 @@ POSTHOOK: query: select * from groupingsets32a POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets32a #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 2147483647 2147483647 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3221225471 3221225471 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3758096383 3758096383 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4026531839 4026531839 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4160749567 4160749567 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4227858431 4227858431 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4261412863 4261412863 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4278190079 4278190079 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4286578687 4286578687 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4290772991 4290772991 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4292870143 4292870143 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4293918719 4293918719 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294443007 4294443007 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294705151 4294705151 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294836223 4294836223 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294901759 4294901759 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294934527 4294934527 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294950911 4294950911 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294959103 4294959103 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294963199 4294963199 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294965247 4294965247 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966271 4294966271 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966783 4294966783 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967039 4294967039 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 4294967167 4294967167 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 4294967231 4294967231 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL 1 4294967263 4294967263 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL 1 4294967279 4294967279 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL 1 4294967287 4294967287 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 3 NULL NULL 1 4294967291 4294967291 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL 1 4294967293 4294967293 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 1 4294967294 4294967294 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL 1 4294967231 4294967231 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL 1 4294967167 4294967167 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967039 4294967039 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966783 4294966783 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294966271 4294966271 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294965247 4294965247 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294963199 4294963199 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294959103 4294959103 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294950911 4294950911 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294934527 4294934527 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294901759 4294901759 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294836223 4294836223 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294705151 4294705151 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294443007 4294443007 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4293918719 4293918719 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4292870143 4292870143 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4290772991 4290772991 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4286578687 4286578687 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4278190079 4278190079 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4261412863 4261412863 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4227858431 4227858431 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4160749567 4160749567 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4026531839 4026531839 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3758096383 3758096383 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 3221225471 3221225471 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 2147483647 2147483647 PREHOOK: query: create table groupingsets33 as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -989,32 +989,6 @@ POSTHOOK: query: select * from groupingsets33 POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets33 #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL NULL 1 @@ -1022,6 +996,32 @@ NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 NULL 1 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 6 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 PREHOOK: query: create table groupingsets33a as select c00,c01,c02,c03,c04,c05,c06,c07,c08,c09, @@ -1134,32 +1134,6 @@ POSTHOOK: query: select * from groupingsets33a POSTHOOK: type: QUERY POSTHOOK: Input: default@groupingsets33a #### A masked pattern was here #### -a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967295 4294967295 -NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 6442450943 6442450943 -NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 7516192767 7516192767 -NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8053063679 8053063679 -NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8321499135 8321499135 -NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8455716863 8455716863 -NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8522825727 8522825727 -NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8556380159 8556380159 -NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8573157375 8573157375 -NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8581545983 8581545983 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8585740287 8585740287 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8587837439 8587837439 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8588886015 8588886015 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589410303 8589410303 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589672447 8589672447 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589803519 8589803519 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589869055 8589869055 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589901823 8589901823 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589918207 8589918207 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589926399 8589926399 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589930495 8589930495 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589932543 8589932543 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589933567 8589933567 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934079 8589934079 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934335 8589934335 -NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 8589934463 8589934463 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 0 NULL NULL NULL NULL NULL NULL 1 8589934527 8589934527 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 NULL NULL NULL NULL NULL 1 8589934559 8589934559 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 2 NULL NULL NULL NULL 1 8589934575 8589934575 @@ -1167,3 +1141,29 @@ NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 4 NULL NULL 1 8589934587 8589934587 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 5 NULL 1 8589934589 8589934589 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 6 1 8589934590 8589934590 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL z NULL NULL NULL NULL NULL NULL NULL 1 8589934463 8589934463 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL y NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934335 8589934335 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL x NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589934079 8589934079 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL w NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589933567 8589933567 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL v NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589932543 8589932543 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL u NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589930495 8589930495 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL t NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589926399 8589926399 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL s NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589918207 8589918207 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL r NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589901823 8589901823 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL q NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589869055 8589869055 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL p NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589803519 8589803519 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL o NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589672447 8589672447 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL n NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8589410303 8589410303 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL m NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8588886015 8588886015 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL l NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8587837439 8587837439 +NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL k NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8585740287 8585740287 +NULL NULL NULL NULL NULL NULL NULL NULL NULL j NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8581545983 8581545983 +NULL NULL NULL NULL NULL NULL NULL NULL i NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8573157375 8573157375 +NULL NULL NULL NULL NULL NULL NULL h NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8556380159 8556380159 +NULL NULL NULL NULL NULL NULL g NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8522825727 8522825727 +NULL NULL NULL NULL NULL f NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8455716863 8455716863 +NULL NULL NULL NULL e NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8321499135 8321499135 +NULL NULL NULL d NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 8053063679 8053063679 +NULL NULL c NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 7516192767 7516192767 +NULL b NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 6442450943 6442450943 +a NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 1 4294967295 4294967295 diff --git ql/src/test/results/clientpositive/llap/hashjoin.q.out ql/src/test/results/clientpositive/llap/hashjoin.q.out new file mode 100644 index 0000000000..b95e69a8b3 --- /dev/null +++ ql/src/test/results/clientpositive/llap/hashjoin.q.out @@ -0,0 +1,586 @@ +PREHOOK: query: CREATE TABLE t_hashjoin_big( + cint int, + cvarchar varchar(50), + cdouble double, + a int +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t_hashjoin_big +POSTHOOK: query: CREATE TABLE t_hashjoin_big( + cint int, + cvarchar varchar(50), + cdouble double, + a int +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t_hashjoin_big +PREHOOK: query: CREATE TABLE t_hashjoin_small( + cint int, + cvarchar varchar(50), + cdouble double +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@t_hashjoin_small +POSTHOOK: query: CREATE TABLE t_hashjoin_small( + cint int, + cvarchar varchar(50), + cdouble double +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@t_hashjoin_small +PREHOOK: query: INSERT INTO t_hashjoin_big VALUES +(5, 'two', 3.0, 1), +(6, 'two', 1.5, 2), +(NULL, NULL, NULL, NULL), +(7, 'eight', 4.2, 3), (7, 'eight', 4.2, 4), (7, 'eight', 4.2, 5), +(5, 'one', 2.8, 6), (5, 'one', 2.8, 7), (5, 'one', 2.8, 8) +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@t_hashjoin_big +POSTHOOK: query: INSERT INTO t_hashjoin_big VALUES +(5, 'two', 3.0, 1), +(6, 'two', 1.5, 2), +(NULL, NULL, NULL, NULL), +(7, 'eight', 4.2, 3), (7, 'eight', 4.2, 4), (7, 'eight', 4.2, 5), +(5, 'one', 2.8, 6), (5, 'one', 2.8, 7), (5, 'one', 2.8, 8) +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@t_hashjoin_big +POSTHOOK: Lineage: t_hashjoin_big.a SCRIPT [] +POSTHOOK: Lineage: t_hashjoin_big.cdouble SCRIPT [] +POSTHOOK: Lineage: t_hashjoin_big.cint SCRIPT [] +POSTHOOK: Lineage: t_hashjoin_big.cvarchar SCRIPT [] +PREHOOK: query: INSERT INTO t_hashjoin_small VALUES +(7, 'two', 1.5), +(5, 'two', 4.2), +(NULL, NULL, NULL), +(5, 'one', 1.1), (5, 'one', 1.1) +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@t_hashjoin_small +POSTHOOK: query: INSERT INTO t_hashjoin_small VALUES +(7, 'two', 1.5), +(5, 'two', 4.2), +(NULL, NULL, NULL), +(5, 'one', 1.1), (5, 'one', 1.1) +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@t_hashjoin_small +POSTHOOK: Lineage: t_hashjoin_small.cdouble SCRIPT [] +POSTHOOK: Lineage: t_hashjoin_small.cint SCRIPT [] +POSTHOOK: Lineage: t_hashjoin_small.cvarchar SCRIPT [] +PREHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: cint is not null (type: boolean) + Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cint is not null (type: boolean) + Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 cint (type: int) + 1 cint (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 + input vertices: + 1 Map 2 + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: z + filterExpr: cint is not null (type: boolean) + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cint is not null (type: boolean) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: cint (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: cint (type: int) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: cvarchar (type: varchar(50)), cdouble (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +7 eight 4.2 3 7 two 1.5 +7 eight 4.2 4 7 two 1.5 +7 eight 4.2 5 7 two 1.5 +5 two 3.0 1 5 two 4.2 +5 two 3.0 1 5 one 1.1 +5 two 3.0 1 5 one 1.1 +5 one 2.8 6 5 two 4.2 +5 one 2.8 6 5 one 1.1 +5 one 2.8 6 5 one 1.1 +5 one 2.8 7 5 two 4.2 +5 one 2.8 7 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 8 5 two 4.2 +5 one 2.8 8 5 one 1.1 +5 one 2.8 8 5 one 1.1 +PREHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: cvarchar is not null (type: boolean) + Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cvarchar is not null (type: boolean) + Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 cvarchar (type: varchar(50)) + 1 cvarchar (type: varchar(50)) + outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 + input vertices: + 1 Map 2 + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 2010 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: z + filterExpr: cvarchar is not null (type: boolean) + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cvarchar is not null (type: boolean) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: cvarchar (type: varchar(50)) + null sort order: z + sort order: + + Map-reduce partition columns: cvarchar (type: varchar(50)) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: cint (type: int), cdouble (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +5 two 3.0 1 7 two 1.5 +5 two 3.0 1 5 two 4.2 +6 two 1.5 2 7 two 1.5 +6 two 1.5 2 5 two 4.2 +5 one 2.8 6 5 one 1.1 +5 one 2.8 6 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 8 5 one 1.1 +5 one 2.8 8 5 one 1.1 +PREHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: (cint is not null and cvarchar is not null) (type: boolean) + Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (cint is not null and cvarchar is not null) (type: boolean) + Statistics: Num rows: 7 Data size: 721 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 cint (type: int), cvarchar (type: varchar(50)) + 1 cint (type: int), cvarchar (type: varchar(50)) + outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 + input vertices: + 1 Map 2 + Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 7 Data size: 1407 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: z + filterExpr: (cint is not null and cvarchar is not null) (type: boolean) + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (cint is not null and cvarchar is not null) (type: boolean) + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: cint (type: int), cvarchar (type: varchar(50)) + null sort order: zz + sort order: ++ + Map-reduce partition columns: cint (type: int), cvarchar (type: varchar(50)) + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: cdouble (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +5 two 3.0 1 5 two 4.2 +5 one 2.8 6 5 one 1.1 +5 one 2.8 6 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 8 5 one 1.1 +5 one 2.8 8 5 one 1.1 +PREHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 cint (type: int), cvarchar (type: varchar(50)) + 1 cint (type: int), cvarchar (type: varchar(50)) + outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 + input vertices: + 1 Map 2 + Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 19 Data size: 3427 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: z + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: cint (type: int), cvarchar (type: varchar(50)) + null sort order: zz + sort order: ++ + Map-reduce partition columns: cint (type: int), cvarchar (type: varchar(50)) + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: cdouble (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM t_hashjoin_big x LEFT OUTER JOIN t_hashjoin_small z ON (x.cint = z.cint AND x.cvarchar = z.cvarchar) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +5 two 3.0 1 5 two 4.2 +6 two 1.5 2 NULL NULL NULL +NULL NULL NULL NULL NULL NULL NULL +7 eight 4.2 3 NULL NULL NULL +7 eight 4.2 4 NULL NULL NULL +7 eight 4.2 5 NULL NULL NULL +5 one 2.8 6 5 one 1.1 +5 one 2.8 6 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 7 5 one 1.1 +5 one 2.8 8 5 one 1.1 +5 one 2.8 8 5 one 1.1 +PREHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: cdouble is not null (type: boolean) + Statistics: Num rows: 9 Data size: 927 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cdouble is not null (type: boolean) + Statistics: Num rows: 8 Data size: 824 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 cdouble (type: double) + 1 cdouble (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col7, _col8, _col9 + input vertices: + 1 Map 2 + Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: varchar(50)), _col2 (type: double), _col3 (type: int), _col7 (type: int), _col8 (type: varchar(50)), _col9 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 8 Data size: 1608 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: z + filterExpr: cdouble is not null (type: boolean) + Statistics: Num rows: 5 Data size: 490 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: cdouble is not null (type: boolean) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: cdouble (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: cdouble (type: double) + Statistics: Num rows: 4 Data size: 392 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: cint (type: int), cvarchar (type: varchar(50)) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) +PREHOOK: type: QUERY +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM t_hashjoin_big x JOIN t_hashjoin_small z ON (x.cdouble = z.cdouble) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Input: default@t_hashjoin_small +#### A masked pattern was here #### +6 two 1.5 2 7 two 1.5 +7 eight 4.2 3 5 two 4.2 +7 eight 4.2 4 5 two 4.2 +7 eight 4.2 5 5 two 4.2 +PREHOOK: query: DROP TABLE t_hashjoin_big +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@t_hashjoin_big +PREHOOK: Output: default@t_hashjoin_big +POSTHOOK: query: DROP TABLE t_hashjoin_big +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@t_hashjoin_big +POSTHOOK: Output: default@t_hashjoin_big +PREHOOK: query: DROP TABLE t_hashjoin_small +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@t_hashjoin_small +PREHOOK: Output: default@t_hashjoin_small +POSTHOOK: query: DROP TABLE t_hashjoin_small +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@t_hashjoin_small +POSTHOOK: Output: default@t_hashjoin_small diff --git ql/src/test/results/clientpositive/llap/having2.q.out ql/src/test/results/clientpositive/llap/having2.q.out new file mode 100644 index 0000000000..dbf0f30c0d --- /dev/null +++ ql/src/test/results/clientpositive/llap/having2.q.out @@ -0,0 +1,692 @@ +PREHOOK: query: CREATE TABLE TestV1_Staples ( + Item_Count INT, + Ship_Priority STRING, + Order_Priority STRING, + Order_Status STRING, + Order_Quantity DOUBLE, + Sales_Total DOUBLE, + Discount DOUBLE, + Tax_Rate DOUBLE, + Ship_Mode STRING, + Fill_Time DOUBLE, + Gross_Profit DOUBLE, + Price DOUBLE, + Ship_Handle_Cost DOUBLE, + Employee_Name STRING, + Employee_Dept STRING, + Manager_Name STRING, + Employee_Yrs_Exp DOUBLE, + Employee_Salary DOUBLE, + Customer_Name STRING, + Customer_State STRING, + Call_Center_Region STRING, + Customer_Balance DOUBLE, + Customer_Segment STRING, + Prod_Type1 STRING, + Prod_Type2 STRING, + Prod_Type3 STRING, + Prod_Type4 STRING, + Product_Name STRING, + Product_Container STRING, + Ship_Promo STRING, + Supplier_Name STRING, + Supplier_Balance DOUBLE, + Supplier_Region STRING, + Supplier_State STRING, + Order_ID STRING, + Order_Year INT, + Order_Month INT, + Order_Day INT, + Order_Date_ STRING, + Order_Quarter STRING, + Product_Base_Margin DOUBLE, + Product_ID STRING, + Receive_Time DOUBLE, + Received_Date_ STRING, + Ship_Date_ STRING, + Ship_Charge DOUBLE, + Total_Cycle_Time DOUBLE, + Product_In_Stock STRING, + PID INT, + Market_Segment STRING + ) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@TestV1_Staples +POSTHOOK: query: CREATE TABLE TestV1_Staples ( + Item_Count INT, + Ship_Priority STRING, + Order_Priority STRING, + Order_Status STRING, + Order_Quantity DOUBLE, + Sales_Total DOUBLE, + Discount DOUBLE, + Tax_Rate DOUBLE, + Ship_Mode STRING, + Fill_Time DOUBLE, + Gross_Profit DOUBLE, + Price DOUBLE, + Ship_Handle_Cost DOUBLE, + Employee_Name STRING, + Employee_Dept STRING, + Manager_Name STRING, + Employee_Yrs_Exp DOUBLE, + Employee_Salary DOUBLE, + Customer_Name STRING, + Customer_State STRING, + Call_Center_Region STRING, + Customer_Balance DOUBLE, + Customer_Segment STRING, + Prod_Type1 STRING, + Prod_Type2 STRING, + Prod_Type3 STRING, + Prod_Type4 STRING, + Product_Name STRING, + Product_Container STRING, + Ship_Promo STRING, + Supplier_Name STRING, + Supplier_Balance DOUBLE, + Supplier_Region STRING, + Supplier_State STRING, + Order_ID STRING, + Order_Year INT, + Order_Month INT, + Order_Day INT, + Order_Date_ STRING, + Order_Quarter STRING, + Product_Base_Margin DOUBLE, + Product_ID STRING, + Receive_Time DOUBLE, + Received_Date_ STRING, + Ship_Date_ STRING, + Ship_Charge DOUBLE, + Total_Cycle_Time DOUBLE, + Product_In_Stock STRING, + PID INT, + Market_Segment STRING + ) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@TestV1_Staples +PREHOOK: query: explain +SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( +(COUNT(s1.discount) <= 822) AND +(SUM(customer_balance) <= 4074689.000000041) +) +PREHOOK: type: QUERY +PREHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( +(COUNT(s1.discount) <= 822) AND +(SUM(customer_balance) <= 4074689.000000041) +) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: order_quantity (type: double), discount (type: double), customer_name (type: string), customer_balance (type: double) + outputColumnNames: order_quantity, discount, customer_name, customer_balance + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(customer_balance), sum(order_quantity), count(discount) + keys: customer_name (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((_col3 <= 822L) and (_col1 <= 4074689.000000041D)) (type: boolean) + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( +(SUM(customer_balance) <= 4074689.000000041) +AND (COUNT(s1.discount) <= 822) +) +PREHOOK: type: QUERY +PREHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT customer_name, SUM(customer_balance), SUM(order_quantity) FROM default.testv1_staples s1 GROUP BY customer_name HAVING ( +(SUM(customer_balance) <= 4074689.000000041) +AND (COUNT(s1.discount) <= 822) +) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: order_quantity (type: double), discount (type: double), customer_name (type: string), customer_balance (type: double) + outputColumnNames: order_quantity, discount, customer_name, customer_balance + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(customer_balance), sum(order_quantity), count(discount) + keys: customer_name (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((_col1 <= 4074689.000000041D) and (_col3 <= 822L)) (type: boolean) + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 208 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +GROUP BY s1.customer_name +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +GROUP BY s1.customer_name +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + filterExpr: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: double), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: s2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(_col2), count(_col4), sum(_col0), count(_col0) + keys: _col1 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: double), _col2 (type: bigint), _col3 (type: double), _col4 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1), sum(VALUE._col2), count(VALUE._col3) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((_col1 <= 4074689.000000041D) and (_col2 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +GROUP BY s1.customer_name, s1.customer_name +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT s1.customer_name FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +GROUP BY s1.customer_name, s1.customer_name +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + filterExpr: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: double), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: s2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(_col2), sum(_col0), count(_col0), count(_col4) + keys: _col1 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2), count(VALUE._col3) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) + outputColumnNames: _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((_col2 <= 4074689.000000041D) and (_col5 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +SELECT distinct s1.customer_name as x, s1.customer_name as y +FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT distinct s1.customer_name as x, s1.customer_name as y +FROM default.testv1_staples s1 join default.src s2 on s1.customer_name = s2.key +HAVING ( +(SUM(s1.customer_balance) <= 4074689.000000041) +AND (AVG(s1.discount) <= 822) +AND (COUNT(s2.value) > 4) +) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@testv1_staples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + filterExpr: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: customer_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: discount (type: double), customer_name (type: string), customer_balance (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 200 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: double), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: s2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: sum(_col2), sum(_col0), count(_col0), count(_col4) + keys: _col1 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 550 Data size: 97900 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), sum(VALUE._col1), count(VALUE._col2), count(VALUE._col3) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: double), _col2 (type: double), _col3 (type: bigint), _col4 (type: bigint) + outputColumnNames: _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 275 Data size: 48950 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((_col2 <= 4074689.000000041D) and (_col5 > 4L) and ((_col3 / _col4) <= 822.0D)) (type: boolean) + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col1 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/hll.q.out ql/src/test/results/clientpositive/llap/hll.q.out similarity index 74% rename from ql/src/test/results/clientpositive/hll.q.out rename to ql/src/test/results/clientpositive/llap/hll.q.out index 9cb85bb61c..0f2e13b1b2 100644 --- ql/src/test/results/clientpositive/hll.q.out +++ ql/src/test/results/clientpositive/llap/hll.q.out @@ -27,45 +27,55 @@ POSTHOOK: Output: default@n #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-0 is a root stage - Stage-1 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-0 - Map Reduce - Map Operator Tree: - TableScan - alias: n - Statistics: Num rows: 500 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: n + Statistics: Num rows: 500 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0) + mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-1 + Stage: Stage-2 Stats Work Basic Stats Work: Column Stats Desc: @@ -131,45 +141,55 @@ POSTHOOK: Output: default@i #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-0 is a root stage - Stage-1 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-0 - Map Reduce - Map Operator Tree: - TableScan - alias: i - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: i + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0) + mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-1 + Stage: Stage-2 Stats Work Basic Stats Work: Column Stats Desc: diff --git ql/src/test/results/clientpositive/implicit_cast1.q.out ql/src/test/results/clientpositive/llap/implicit_cast1.q.out similarity index 56% rename from ql/src/test/results/clientpositive/implicit_cast1.q.out rename to ql/src/test/results/clientpositive/llap/implicit_cast1.q.out index 1b2d13ce34..0bb992616b 100644 --- ql/src/test/results/clientpositive/implicit_cast1.q.out +++ ql/src/test/results/clientpositive/llap/implicit_cast1.q.out @@ -21,38 +21,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@implicit_test1 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: implicit_test1 - filterExpr: (a <> 0L) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (a <> 0L) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: bigint), b (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: implicit_test1 + filterExpr: (a <> 0L) (type: boolean) + Filter Operator + predicate: (a <> 0L) (type: boolean) + Select Operator + expressions: a (type: bigint), b (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: SELECT implicit_test1.* FROM implicit_test1 diff --git ql/src/test/results/clientpositive/llap/implicit_cast_during_insert.q.out ql/src/test/results/clientpositive/llap/implicit_cast_during_insert.q.out new file mode 100644 index 0000000000..c62a816b7a --- /dev/null +++ ql/src/test/results/clientpositive/llap/implicit_cast_during_insert.q.out @@ -0,0 +1,154 @@ +PREHOOK: query: create table implicit_cast_during_insert (c1 int, c2 string) + partitioned by (p1 string) stored as orc +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@implicit_cast_during_insert +POSTHOOK: query: create table implicit_cast_during_insert (c1 int, c2 string) + partitioned by (p1 string) stored as orc +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@implicit_cast_during_insert +PREHOOK: query: explain +insert overwrite table implicit_cast_during_insert partition (p1) + select key, value, key key1 from (select * from src where key in (0,1)) q + distribute by key1 sort by key1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@implicit_cast_during_insert +POSTHOOK: query: explain +insert overwrite table implicit_cast_during_insert partition (p1) + select key, value, key key1 from (select * from src where key in (0,1)) q + distribute by key1 sort by key1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (key) IN (0, 1) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key) IN (0, 1) (type: boolean) + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string), key (type: string) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: UDFToInteger(KEY.reducesinkkey0) (type: int), VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.implicit_cast_during_insert + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: c1, c2, p1 + Statistics: Num rows: 3 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') + keys: p1 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 967 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 967 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 967 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + p1 + replace: true + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.implicit_cast_during_insert + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2 + Column Types: int, string + Table: default.implicit_cast_during_insert + +PREHOOK: query: insert overwrite table implicit_cast_during_insert partition (p1) + select key, value, key key1 from (select * from src where key in (0,1)) q + distribute by key1 sort by key1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@implicit_cast_during_insert +POSTHOOK: query: insert overwrite table implicit_cast_during_insert partition (p1) + select key, value, key key1 from (select * from src where key in (0,1)) q + distribute by key1 sort by key1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@implicit_cast_during_insert@p1=0 +POSTHOOK: Lineage: implicit_cast_during_insert PARTITION(p1=0).c1 EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: implicit_cast_during_insert PARTITION(p1=0).c2 SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from implicit_cast_during_insert +PREHOOK: type: QUERY +PREHOOK: Input: default@implicit_cast_during_insert +PREHOOK: Input: default@implicit_cast_during_insert@p1=0 +#### A masked pattern was here #### +POSTHOOK: query: select * from implicit_cast_during_insert +POSTHOOK: type: QUERY +POSTHOOK: Input: default@implicit_cast_during_insert +POSTHOOK: Input: default@implicit_cast_during_insert@p1=0 +#### A masked pattern was here #### +0 val_0 0 +0 val_0 0 +0 val_0 0 +PREHOOK: query: drop table implicit_cast_during_insert +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@implicit_cast_during_insert +PREHOOK: Output: default@implicit_cast_during_insert +POSTHOOK: query: drop table implicit_cast_during_insert +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@implicit_cast_during_insert +POSTHOOK: Output: default@implicit_cast_during_insert diff --git ql/src/test/results/clientpositive/implicit_decimal.q.out ql/src/test/results/clientpositive/llap/implicit_decimal.q.out similarity index 66% rename from ql/src/test/results/clientpositive/implicit_decimal.q.out rename to ql/src/test/results/clientpositive/llap/implicit_decimal.q.out index 6d7b28a730..83a9465e66 100644 --- ql/src/test/results/clientpositive/implicit_decimal.q.out +++ ql/src/test/results/clientpositive/llap/implicit_decimal.q.out @@ -35,27 +35,31 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: decimal_test - filterExpr: (dc = 4327269606205.029297) (type: boolean) - Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (dc = 4327269606205.029297) (type: boolean) - Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 4327269606205.029297 (type: decimal(38,18)) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Tez +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: decimal_test + filterExpr: (dc = 4327269606205.029297) (type: boolean) Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized + Filter Operator + predicate: (dc = 4327269606205.029297) (type: boolean) + Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 4327269606205.029297 (type: decimal(38,18)) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/in_typecheck_char.q.out ql/src/test/results/clientpositive/llap/in_typecheck_char.q.out new file mode 100644 index 0000000000..65b9548389 --- /dev/null +++ ql/src/test/results/clientpositive/llap/in_typecheck_char.q.out @@ -0,0 +1,461 @@ +PREHOOK: query: create table ax(s char(1),t char(10)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@ax +POSTHOOK: query: create table ax(s char(1),t char(10)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@ax +PREHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@ax +POSTHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@ax +POSTHOOK: Lineage: ax.s SCRIPT [] +POSTHOOK: Lineage: ax.t SCRIPT [] +PREHOOK: query: explain +select 'expected 2',count(*) from ax where s = 'a' and t = 'a' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain +select 'expected 2',count(*) from ax where s = 'a' and t = 'a' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: ((s = 'a') and (t = 'a ')) (type: boolean) + Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((s = 'a') and (t = 'a ')) (type: boolean) + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'expected 2' (type: string), _col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 'expected 2',count(*) from ax where s = 'a' and t = 'a' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 2',count(*) from ax where s = 'a' and t = 'a' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 2 2 +PREHOOK: query: explain +select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain +select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) + Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'expected 3' (type: string), _col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 3',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 3 3 +PREHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 2 2 +PREHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 2',count(*) from ax where t = 'a ' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 2 2 +PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 0 0 +PREHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),(null, 'bb')) +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),(null, 'bb')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 2 2 +PREHOOK: query: explain select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: (((CAST( s AS STRING) = 'a') and (CAST( t AS STRING) = 'a')) or (null and (CAST( t AS STRING) = 'bb'))) is null (type: boolean) + Statistics: Num rows: 3 Data size: 510 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((CAST( s AS STRING) = 'a') and (CAST( t AS STRING) = 'a')) or (null and (CAST( t AS STRING) = 'bb'))) is null (type: boolean) + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'expected 1' (type: string), _col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 1',count(*) from ax where ((s,t) in (('a','a'),(null, 'bb'))) is null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 1 1 +PREHOOK: query: explain cbo select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain cbo select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +CBO PLAN: +HiveAggregate(group=[{}], agg#0=[count()]) + HiveFilter(condition=[IN($1, _UTF-16LE'a ', _UTF-16LE'bb ', _UTF-16LE'aa ', _UTF-16LE'bbb ', _UTF-16LE'ab ', _UTF-16LE'ba ', _UTF-16LE'aaa ', _UTF-16LE'bbb ', _UTF-16LE'abc ', _UTF-16LE'bc ', _UTF-16LE'ac ', _UTF-16LE'bca ', _UTF-16LE'cab ', _UTF-16LE'cb ', _UTF-16LE'ca ', _UTF-16LE'cbc ', _UTF-16LE'z ')]) + HiveTableScan(table=[[default, ax]], table:alias=[ax]) + +PREHOOK: query: explain select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: (t) IN ('a ', 'bb ', 'aa ', 'bbb ', 'ab ', 'ba ', 'aaa ', 'bbb ', 'abc ', 'bc ', 'ac ', 'bca ', 'cab ', 'cb ', 'ca ', 'cbc ', 'z ') (type: boolean) + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (t) IN ('a ', 'bb ', 'aa ', 'bbb ', 'ab ', 'ba ', 'aaa ', 'bbb ', 'abc ', 'bc ', 'ac ', 'bca ', 'cab ', 'cb ', 'ca ', 'cbc ', 'z ') (type: boolean) + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain cbo select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain cbo select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +CBO PLAN: +HiveAggregate(group=[{}], agg#0=[count()]) + HiveFilter(condition=[OR(=($1, _UTF-16LE'a '), =($1, _UTF-16LE'bb '), =($1, _UTF-16LE'aa '), =($1, _UTF-16LE'bbb '), =($1, _UTF-16LE'ab '), =($1, _UTF-16LE'ba '), =($1, _UTF-16LE'aaa '), =($1, _UTF-16LE'abc '), =($1, _UTF-16LE'bc '), =($1, _UTF-16LE'ac '), =($1, _UTF-16LE'bca '), =($1, _UTF-16LE'cab '), =($1, _UTF-16LE'cb '), =($1, _UTF-16LE'ca '), =($1, _UTF-16LE'cbc '), =($1, _UTF-16LE'z '))]) + HiveTableScan(table=[[default, ax]], table:alias=[ax]) + +PREHOOK: query: explain select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain select count(*) from ax where t in +('a','bb','aa','bbb','ab','ba','aaa','bbb','abc','bc','ac','bca','cab','cb','ca','cbc','z') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: ((t = 'a ') or (t = 'bb ') or (t = 'aa ') or (t = 'bbb ') or (t = 'ab ') or (t = 'ba ') or (t = 'aaa ') or (t = 'abc ') or (t = 'bc ') or (t = 'ac ') or (t = 'bca ') or (t = 'cab ') or (t = 'cb ') or (t = 'ca ') or (t = 'cbc ') or (t = 'z ')) (type: boolean) + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((t = 'a ') or (t = 'bb ') or (t = 'aa ') or (t = 'bbb ') or (t = 'ab ') or (t = 'ba ') or (t = 'aaa ') or (t = 'abc ') or (t = 'bc ') or (t = 'ac ') or (t = 'bca ') or (t = 'cab ') or (t = 'cb ') or (t = 'ca ') or (t = 'cbc ') or (t = 'z ')) (type: boolean) + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 3 Data size: 255 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/in_typecheck_mixed.q.out ql/src/test/results/clientpositive/llap/in_typecheck_mixed.q.out similarity index 64% rename from ql/src/test/results/clientpositive/in_typecheck_mixed.q.out rename to ql/src/test/results/clientpositive/llap/in_typecheck_mixed.q.out index a7f89d1b65..2ee05ec511 100644 --- ql/src/test/results/clientpositive/in_typecheck_mixed.q.out +++ ql/src/test/results/clientpositive/llap/in_typecheck_mixed.q.out @@ -26,38 +26,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@t #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t - filterExpr: ((UDFToDouble(a)) IN (1.0D, 2.0D) or (a = 'x')) (type: boolean) - Statistics: Num rows: 3 Data size: 258 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(a)) IN (1.0D, 2.0D) or (a = 'x')) (type: boolean) - Statistics: Num rows: 2 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: a (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 172 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: t + filterExpr: ((UDFToDouble(a)) IN (1.0D, 2.0D) or (a = 'x')) (type: boolean) + Filter Operator + predicate: ((UDFToDouble(a)) IN (1.0D, 2.0D) or (a = 'x')) (type: boolean) + Select Operator + expressions: a (type: string) + outputColumnNames: _col0 + ListSink PREHOOK: query: select * from t where a in (1.0,'x',2) PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/in_typecheck_pointlook.q.out ql/src/test/results/clientpositive/llap/in_typecheck_pointlook.q.out new file mode 100644 index 0000000000..2f39378084 --- /dev/null +++ ql/src/test/results/clientpositive/llap/in_typecheck_pointlook.q.out @@ -0,0 +1,204 @@ +PREHOOK: query: create table customer_demographics (cd_marital_status char(1), cd_education_status char(20)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@customer_demographics +POSTHOOK: query: create table customer_demographics (cd_marital_status char(1), cd_education_status char(20)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@customer_demographics +PREHOOK: query: insert into customer_demographics values +('M','Unknown'), +('W','Advanced Degree'), +('W','Advanced Degree '), +('W',' Advanced Degree') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@customer_demographics +POSTHOOK: query: insert into customer_demographics values +('M','Unknown'), +('W','Advanced Degree'), +('W','Advanced Degree '), +('W',' Advanced Degree') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@customer_demographics +POSTHOOK: Lineage: customer_demographics.cd_education_status SCRIPT [] +POSTHOOK: Lineage: customer_demographics.cd_marital_status SCRIPT [] +PREHOOK: query: explain +select count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +PREHOOK: type: QUERY +PREHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +POSTHOOK: query: explain +select count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: customer_demographics + filterExpr: (((cd_marital_status = 'M') and (cd_education_status = 'Unknown ')) or ((cd_marital_status = 'W') and (cd_education_status = 'Advanced Degree '))) (type: boolean) + Statistics: Num rows: 4 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((cd_marital_status = 'M') and (cd_education_status = 'Unknown ')) or ((cd_marital_status = 'W') and (cd_education_status = 'Advanced Degree '))) (type: boolean) + Statistics: Num rows: 2 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 2 Data size: 364 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select '3 is expected:',count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +PREHOOK: type: QUERY +PREHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +POSTHOOK: query: select '3 is expected:',count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +3 is expected: 3 +PREHOOK: query: explain +select count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +PREHOOK: type: QUERY +PREHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +POSTHOOK: query: explain +select count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: customer_demographics + filterExpr: (struct(cd_marital_status,cd_education_status)) IN (const struct('M','Unknown '), const struct('W','Advanced Degree ')) (type: boolean) + Statistics: Num rows: 4 Data size: 728 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (struct(cd_marital_status,cd_education_status)) IN (const struct('M','Unknown '), const struct('W','Advanced Degree ')) (type: boolean) + Statistics: Num rows: 1 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 182 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select '3 is expected:',count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +PREHOOK: type: QUERY +PREHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +POSTHOOK: query: select '3 is expected:',count(1) +from customer_demographics +where ( (cd_marital_status = 'M' and cd_education_status = 'Unknown') +or (cd_marital_status = 'W' and cd_education_status = 'Advanced Degree')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@customer_demographics +#### A masked pattern was here #### +3 is expected: 3 diff --git ql/src/test/results/clientpositive/llap/in_typecheck_varchar.q.out ql/src/test/results/clientpositive/llap/in_typecheck_varchar.q.out new file mode 100644 index 0000000000..1dcf2c5b60 --- /dev/null +++ ql/src/test/results/clientpositive/llap/in_typecheck_varchar.q.out @@ -0,0 +1,207 @@ +PREHOOK: query: create table ax(s varchar(1),t varchar(10)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@ax +POSTHOOK: query: create table ax(s varchar(1),t varchar(10)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@ax +PREHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@ax +POSTHOOK: query: insert into ax values ('a','a'),('a','a '),('b','bb') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@ax +POSTHOOK: Lineage: ax.s SCRIPT [] +POSTHOOK: Lineage: ax.t SCRIPT [] +PREHOOK: query: explain +select 'expected 1',count(*) from ax where s = 'a' and t = 'a' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain +select 'expected 1',count(*) from ax where s = 'a' and t = 'a' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: ((s = 'a') and (t = 'a')) (type: boolean) + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((s = 'a') and (t = 'a')) (type: boolean) + Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'expected 1' (type: string), _col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 'expected 1',count(*) from ax where s = 'a' and t = 'a' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 1',count(*) from ax where s = 'a' and t = 'a' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 1 1 +PREHOOK: query: explain +select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: explain +select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: ax + filterExpr: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) + Statistics: Num rows: 3 Data size: 513 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (struct(CAST( s AS STRING),CAST( t AS STRING))) IN (const struct('a','a'), const struct('b','bb')) (type: boolean) + Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 171 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'expected 2' (type: string), _col0 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 102 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 2',count(*) from ax where (s,t) in (('a','a'),('b','bb')) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 2 2 +PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 0 0 +PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a ' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 0 0 +PREHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' +PREHOOK: type: QUERY +PREHOOK: Input: default@ax +#### A masked pattern was here #### +POSTHOOK: query: select 'expected 0',count(*) from ax where t = 'a d' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ax +#### A masked pattern was here #### +expected 0 0 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_convert_join.q.out similarity index 93% rename from ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_convert_join.q.out index 5cc876eeaf..7820705680 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_convert_join.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_convert_join.q.out @@ -60,9 +60,6 @@ SELECT a.key, b.value FROM src a JOIN src b ON a.key = b.key PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@test_table_n11@part=1 -Hive Runtime Error: Map local work exhausted memory -FAILED: Execution Error, return code 3 from org.apache.hadoop.hive.ql.exec.mr.MapredLocalTask -ATTEMPT: Execute BackupTask: org.apache.hadoop.hive.ql.exec.mr.MapRedTask POSTHOOK: query: INSERT OVERWRITE TABLE test_table_n11 PARTITION (part = '1') SELECT a.key, b.value FROM src a JOIN src b ON a.key = b.key POSTHOOK: type: QUERY @@ -102,8 +99,8 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_dyn_part.q.out similarity index 81% rename from ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_dyn_part.q.out index 02c40a24a7..dc5ce6b965 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_dyn_part.q.out @@ -156,10 +156,10 @@ Table: test_table_n8 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1173 - totalSize 1342 + totalSize 1427 #### A masked pattern was here #### # Storage Information @@ -194,10 +194,10 @@ Table: test_table_n8 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1173 - totalSize 1381 + totalSize 1429 #### A masked pattern was here #### # Storage Information @@ -257,10 +257,10 @@ Table: test_table_n8 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 155 rawDataSize 586 - totalSize 722 + totalSize 796 #### A masked pattern was here #### # Storage Information @@ -295,10 +295,10 @@ Table: test_table_n8 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 154 rawDataSize 591 - totalSize 741 + totalSize 798 #### A masked pattern was here #### # Storage Information @@ -419,87 +419,107 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0, Stage-8 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - Stage-8 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.684 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string), if(((UDFToDouble(_col0) % 100.0D) = 0.0D), '11', '12') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 112812 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 316 Data size: 112812 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.test_table_n8 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 316 Data size: 142516 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string), if(((UDFToDouble(_col0) % 100.0D) = 0.0D), '11', '12') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 112812 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 316 Data size: 112812 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.test_table_n8 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 316 Data size: 142516 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-7 - Conditional Operator + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -514,7 +534,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.test_table_n8 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -522,57 +542,6 @@ STAGE PLANS: Column Types: string, string Table: default.test_table_n8 - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-8 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE test_table_n8 PARTITION (ds = '2008-04-08', hr) SELECT key, value, IF (key % 100 == 0, '11', '12') FROM (SELECT key, COUNT(*) AS value FROM srcpart @@ -658,10 +627,10 @@ Table: test_table_n8 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 305 rawDataSize 1163 - totalSize 1347 + totalSize 1428 #### A masked pattern was here #### # Storage Information diff --git ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_grouping_operators.q.out similarity index 67% rename from ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_grouping_operators.q.out index 8a861a5995..0b6952b861 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_grouping_operators.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_grouping_operators.q.out @@ -26,73 +26,107 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@test_table_out@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), value (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 750 Data size: 145500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 750 Data size: 145500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - keys: key (type: string), value (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 750 Data size: 145500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 750 Data size: 145500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 750 Data size: 145500 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 750 Data size: 271500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 750 Data size: 271500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) - outputColumnNames: key, value, agg, part - Statistics: Num rows: 750 Data size: 335250 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 750 Data size: 271500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 750 Data size: 271500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) + outputColumnNames: key, value, agg, part + Statistics: Num rows: 750 Data size: 335250 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') + keys: part (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') - keys: part (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -106,7 +140,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_out - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -114,37 +148,6 @@ STAGE PLANS: Column Types: string, string, string Table: default.test_table_out - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: SELECT key, value, count(1) FROM src GROUP BY ROLLUP (key, value) PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -153,624 +156,624 @@ POSTHOOK: query: SELECT key, value, count(1) FROM src GROUP BY ROLLUP (key, valu POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### -0 val_0 3 0 NULL 3 -10 val_10 1 +0 val_0 3 10 NULL 1 -100 val_100 2 +10 val_10 1 100 NULL 2 -103 val_103 2 +100 val_100 2 103 NULL 2 -104 val_104 2 +103 val_103 2 104 NULL 2 -105 val_105 1 +104 val_104 2 105 NULL 1 -11 val_11 1 +105 val_105 1 11 NULL 1 -111 val_111 1 +11 val_11 1 111 NULL 1 -113 val_113 2 +111 val_111 1 113 NULL 2 -114 val_114 1 +113 val_113 2 114 NULL 1 -116 val_116 1 +114 val_114 1 116 NULL 1 -118 val_118 2 +116 val_116 1 118 NULL 2 -119 val_119 3 +118 val_118 2 119 NULL 3 -12 val_12 2 +119 val_119 3 12 NULL 2 -120 val_120 2 +12 val_12 2 120 NULL 2 -125 val_125 2 +120 val_120 2 125 NULL 2 -126 val_126 1 +125 val_125 2 126 NULL 1 -128 val_128 3 +126 val_126 1 128 NULL 3 -129 val_129 2 +128 val_128 3 129 NULL 2 -131 val_131 1 +129 val_129 2 131 NULL 1 -133 val_133 1 +131 val_131 1 133 NULL 1 -134 val_134 2 +133 val_133 1 134 NULL 2 -136 val_136 1 +134 val_134 2 136 NULL 1 -137 val_137 2 +136 val_136 1 137 NULL 2 -138 val_138 4 +137 val_137 2 138 NULL 4 -143 val_143 1 +138 val_138 4 143 NULL 1 -145 val_145 1 +143 val_143 1 145 NULL 1 -146 val_146 2 +145 val_145 1 146 NULL 2 -149 val_149 2 +146 val_146 2 149 NULL 2 -15 val_15 2 +149 val_149 2 15 NULL 2 -150 val_150 1 +15 val_15 2 150 NULL 1 -152 val_152 2 +150 val_150 1 152 NULL 2 -153 val_153 1 +152 val_152 2 153 NULL 1 -155 val_155 1 +153 val_153 1 155 NULL 1 -156 val_156 1 +155 val_155 1 156 NULL 1 -157 val_157 1 +156 val_156 1 157 NULL 1 -158 val_158 1 +157 val_157 1 158 NULL 1 -160 val_160 1 +158 val_158 1 160 NULL 1 -162 val_162 1 +160 val_160 1 162 NULL 1 -163 val_163 1 +162 val_162 1 163 NULL 1 -164 val_164 2 +163 val_163 1 164 NULL 2 -165 val_165 2 +164 val_164 2 165 NULL 2 -166 val_166 1 +165 val_165 2 166 NULL 1 -167 val_167 3 +166 val_166 1 167 NULL 3 -168 val_168 1 +167 val_167 3 168 NULL 1 -169 val_169 4 +168 val_168 1 169 NULL 4 -17 val_17 1 +169 val_169 4 17 NULL 1 -170 val_170 1 +17 val_17 1 170 NULL 1 -172 val_172 2 +170 val_170 1 172 NULL 2 -174 val_174 2 +172 val_172 2 174 NULL 2 -175 val_175 2 +174 val_174 2 175 NULL 2 -176 val_176 2 +175 val_175 2 176 NULL 2 -177 val_177 1 +176 val_176 2 177 NULL 1 -178 val_178 1 +177 val_177 1 178 NULL 1 -179 val_179 2 +178 val_178 1 179 NULL 2 -18 val_18 2 +179 val_179 2 18 NULL 2 -180 val_180 1 +18 val_18 2 180 NULL 1 -181 val_181 1 +180 val_180 1 181 NULL 1 -183 val_183 1 +181 val_181 1 183 NULL 1 -186 val_186 1 +183 val_183 1 186 NULL 1 -187 val_187 3 +186 val_186 1 187 NULL 3 -189 val_189 1 +187 val_187 3 189 NULL 1 -19 val_19 1 +189 val_189 1 19 NULL 1 -190 val_190 1 +19 val_19 1 190 NULL 1 -191 val_191 2 +190 val_190 1 191 NULL 2 -192 val_192 1 +191 val_191 2 192 NULL 1 -193 val_193 3 +192 val_192 1 193 NULL 3 -194 val_194 1 +193 val_193 3 194 NULL 1 -195 val_195 2 +194 val_194 1 195 NULL 2 -196 val_196 1 +195 val_195 2 196 NULL 1 -197 val_197 2 +196 val_196 1 197 NULL 2 -199 val_199 3 +197 val_197 2 199 NULL 3 -2 val_2 1 +199 val_199 3 2 NULL 1 -20 val_20 1 +2 val_2 1 20 NULL 1 -200 val_200 2 +20 val_20 1 200 NULL 2 -201 val_201 1 +200 val_200 2 201 NULL 1 -202 val_202 1 +201 val_201 1 202 NULL 1 -203 val_203 2 +202 val_202 1 203 NULL 2 -205 val_205 2 +203 val_203 2 205 NULL 2 -207 val_207 2 +205 val_205 2 207 NULL 2 -208 val_208 3 +207 val_207 2 208 NULL 3 -209 val_209 2 +208 val_208 3 209 NULL 2 -213 val_213 2 +209 val_209 2 213 NULL 2 -214 val_214 1 +213 val_213 2 214 NULL 1 -216 val_216 2 +214 val_214 1 216 NULL 2 -217 val_217 2 +216 val_216 2 217 NULL 2 -218 val_218 1 +217 val_217 2 218 NULL 1 -219 val_219 2 +218 val_218 1 219 NULL 2 -221 val_221 2 +219 val_219 2 221 NULL 2 -222 val_222 1 +221 val_221 2 222 NULL 1 -223 val_223 2 +222 val_222 1 223 NULL 2 -224 val_224 2 +223 val_223 2 224 NULL 2 -226 val_226 1 +224 val_224 2 226 NULL 1 -228 val_228 1 +226 val_226 1 228 NULL 1 -229 val_229 2 +228 val_228 1 229 NULL 2 -230 val_230 5 +229 val_229 2 230 NULL 5 -233 val_233 2 +230 val_230 5 233 NULL 2 -235 val_235 1 +233 val_233 2 235 NULL 1 -237 val_237 2 +235 val_235 1 237 NULL 2 -238 val_238 2 +237 val_237 2 238 NULL 2 -239 val_239 2 +238 val_238 2 239 NULL 2 -24 val_24 2 +239 val_239 2 24 NULL 2 -241 val_241 1 +24 val_24 2 241 NULL 1 -242 val_242 2 +241 val_241 1 242 NULL 2 -244 val_244 1 +242 val_242 2 244 NULL 1 -247 val_247 1 +244 val_244 1 247 NULL 1 -248 val_248 1 +247 val_247 1 248 NULL 1 -249 val_249 1 +248 val_248 1 249 NULL 1 -252 val_252 1 +249 val_249 1 252 NULL 1 -255 val_255 2 +252 val_252 1 255 NULL 2 -256 val_256 2 +255 val_255 2 256 NULL 2 -257 val_257 1 +256 val_256 2 257 NULL 1 -258 val_258 1 +257 val_257 1 258 NULL 1 -26 val_26 2 +258 val_258 1 26 NULL 2 -260 val_260 1 +26 val_26 2 260 NULL 1 -262 val_262 1 +260 val_260 1 262 NULL 1 -263 val_263 1 +262 val_262 1 263 NULL 1 -265 val_265 2 +263 val_263 1 265 NULL 2 -266 val_266 1 +265 val_265 2 266 NULL 1 -27 val_27 1 +266 val_266 1 27 NULL 1 -272 val_272 2 +27 val_27 1 272 NULL 2 -273 val_273 3 +272 val_272 2 273 NULL 3 -274 val_274 1 +273 val_273 3 274 NULL 1 -275 val_275 1 +274 val_274 1 275 NULL 1 -277 val_277 4 +275 val_275 1 277 NULL 4 -278 val_278 2 +277 val_277 4 278 NULL 2 -28 val_28 1 +278 val_278 2 28 NULL 1 -280 val_280 2 +28 val_28 1 280 NULL 2 -281 val_281 2 +280 val_280 2 281 NULL 2 -282 val_282 2 +281 val_281 2 282 NULL 2 -283 val_283 1 +282 val_282 2 283 NULL 1 -284 val_284 1 +283 val_283 1 284 NULL 1 -285 val_285 1 +284 val_284 1 285 NULL 1 -286 val_286 1 +285 val_285 1 286 NULL 1 -287 val_287 1 +286 val_286 1 287 NULL 1 -288 val_288 2 +287 val_287 1 288 NULL 2 -289 val_289 1 +288 val_288 2 289 NULL 1 -291 val_291 1 +289 val_289 1 291 NULL 1 -292 val_292 1 +291 val_291 1 292 NULL 1 -296 val_296 1 +292 val_292 1 296 NULL 1 -298 val_298 3 +296 val_296 1 298 NULL 3 -30 val_30 1 +298 val_298 3 30 NULL 1 -302 val_302 1 +30 val_30 1 302 NULL 1 -305 val_305 1 +302 val_302 1 305 NULL 1 -306 val_306 1 +305 val_305 1 306 NULL 1 -307 val_307 2 +306 val_306 1 307 NULL 2 -308 val_308 1 +307 val_307 2 308 NULL 1 -309 val_309 2 +308 val_308 1 309 NULL 2 -310 val_310 1 +309 val_309 2 310 NULL 1 -311 val_311 3 +310 val_310 1 311 NULL 3 -315 val_315 1 +311 val_311 3 315 NULL 1 -316 val_316 3 +315 val_315 1 316 NULL 3 -317 val_317 2 +316 val_316 3 317 NULL 2 -318 val_318 3 +317 val_317 2 318 NULL 3 -321 val_321 2 +318 val_318 3 321 NULL 2 -322 val_322 2 +321 val_321 2 322 NULL 2 -323 val_323 1 +322 val_322 2 323 NULL 1 -325 val_325 2 +323 val_323 1 325 NULL 2 -327 val_327 3 +325 val_325 2 327 NULL 3 -33 val_33 1 +327 val_327 3 33 NULL 1 -331 val_331 2 +33 val_33 1 331 NULL 2 -332 val_332 1 +331 val_331 2 332 NULL 1 -333 val_333 2 +332 val_332 1 333 NULL 2 -335 val_335 1 +333 val_333 2 335 NULL 1 -336 val_336 1 +335 val_335 1 336 NULL 1 -338 val_338 1 +336 val_336 1 338 NULL 1 -339 val_339 1 +338 val_338 1 339 NULL 1 -34 val_34 1 +339 val_339 1 34 NULL 1 -341 val_341 1 +34 val_34 1 341 NULL 1 -342 val_342 2 +341 val_341 1 342 NULL 2 -344 val_344 2 +342 val_342 2 344 NULL 2 -345 val_345 1 +344 val_344 2 345 NULL 1 -348 val_348 5 +345 val_345 1 348 NULL 5 -35 val_35 3 +348 val_348 5 35 NULL 3 -351 val_351 1 +35 val_35 3 351 NULL 1 -353 val_353 2 +351 val_351 1 353 NULL 2 -356 val_356 1 +353 val_353 2 356 NULL 1 -360 val_360 1 +356 val_356 1 360 NULL 1 -362 val_362 1 +360 val_360 1 362 NULL 1 -364 val_364 1 +362 val_362 1 364 NULL 1 -365 val_365 1 +364 val_364 1 365 NULL 1 -366 val_366 1 +365 val_365 1 366 NULL 1 -367 val_367 2 +366 val_366 1 367 NULL 2 -368 val_368 1 +367 val_367 2 368 NULL 1 -369 val_369 3 +368 val_368 1 369 NULL 3 -37 val_37 2 +369 val_369 3 37 NULL 2 -373 val_373 1 +37 val_37 2 373 NULL 1 -374 val_374 1 +373 val_373 1 374 NULL 1 -375 val_375 1 +374 val_374 1 375 NULL 1 -377 val_377 1 +375 val_375 1 377 NULL 1 -378 val_378 1 +377 val_377 1 378 NULL 1 -379 val_379 1 +378 val_378 1 379 NULL 1 -382 val_382 2 +379 val_379 1 382 NULL 2 -384 val_384 3 +382 val_382 2 384 NULL 3 -386 val_386 1 +384 val_384 3 386 NULL 1 -389 val_389 1 +386 val_386 1 389 NULL 1 -392 val_392 1 +389 val_389 1 392 NULL 1 -393 val_393 1 +392 val_392 1 393 NULL 1 -394 val_394 1 +393 val_393 1 394 NULL 1 -395 val_395 2 +394 val_394 1 395 NULL 2 -396 val_396 3 +395 val_395 2 396 NULL 3 -397 val_397 2 +396 val_396 3 397 NULL 2 -399 val_399 2 +397 val_397 2 399 NULL 2 -4 val_4 1 +399 val_399 2 4 NULL 1 -400 val_400 1 +4 val_4 1 400 NULL 1 -401 val_401 5 +400 val_400 1 401 NULL 5 -402 val_402 1 +401 val_401 5 402 NULL 1 -403 val_403 3 +402 val_402 1 403 NULL 3 -404 val_404 2 +403 val_403 3 404 NULL 2 -406 val_406 4 +404 val_404 2 406 NULL 4 -407 val_407 1 +406 val_406 4 407 NULL 1 -409 val_409 3 +407 val_407 1 409 NULL 3 -41 val_41 1 +409 val_409 3 41 NULL 1 -411 val_411 1 +41 val_41 1 411 NULL 1 -413 val_413 2 +411 val_411 1 413 NULL 2 -414 val_414 2 +413 val_413 2 414 NULL 2 -417 val_417 3 +414 val_414 2 417 NULL 3 -418 val_418 1 +417 val_417 3 418 NULL 1 -419 val_419 1 +418 val_418 1 419 NULL 1 -42 val_42 2 +419 val_419 1 42 NULL 2 -421 val_421 1 +42 val_42 2 421 NULL 1 -424 val_424 2 +421 val_421 1 424 NULL 2 -427 val_427 1 +424 val_424 2 427 NULL 1 -429 val_429 2 +427 val_427 1 429 NULL 2 -43 val_43 1 +429 val_429 2 43 NULL 1 -430 val_430 3 +43 val_43 1 430 NULL 3 -431 val_431 3 +430 val_430 3 431 NULL 3 -432 val_432 1 +431 val_431 3 432 NULL 1 -435 val_435 1 +432 val_432 1 435 NULL 1 -436 val_436 1 +435 val_435 1 436 NULL 1 -437 val_437 1 +436 val_436 1 437 NULL 1 -438 val_438 3 +437 val_437 1 438 NULL 3 -439 val_439 2 +438 val_438 3 439 NULL 2 -44 val_44 1 +439 val_439 2 44 NULL 1 -443 val_443 1 +44 val_44 1 443 NULL 1 -444 val_444 1 +443 val_443 1 444 NULL 1 -446 val_446 1 +444 val_444 1 446 NULL 1 -448 val_448 1 +446 val_446 1 448 NULL 1 -449 val_449 1 +448 val_448 1 449 NULL 1 -452 val_452 1 +449 val_449 1 452 NULL 1 -453 val_453 1 +452 val_452 1 453 NULL 1 -454 val_454 3 +453 val_453 1 454 NULL 3 -455 val_455 1 +454 val_454 3 455 NULL 1 -457 val_457 1 +455 val_455 1 457 NULL 1 -458 val_458 2 +457 val_457 1 458 NULL 2 -459 val_459 2 +458 val_458 2 459 NULL 2 -460 val_460 1 +459 val_459 2 460 NULL 1 -462 val_462 2 +460 val_460 1 462 NULL 2 -463 val_463 2 +462 val_462 2 463 NULL 2 -466 val_466 3 +463 val_463 2 466 NULL 3 -467 val_467 1 +466 val_466 3 467 NULL 1 -468 val_468 4 +467 val_467 1 468 NULL 4 -469 val_469 5 +468 val_468 4 469 NULL 5 -47 val_47 1 +469 val_469 5 47 NULL 1 -470 val_470 1 +47 val_47 1 470 NULL 1 -472 val_472 1 +470 val_470 1 472 NULL 1 -475 val_475 1 +472 val_472 1 475 NULL 1 -477 val_477 1 +475 val_475 1 477 NULL 1 -478 val_478 2 +477 val_477 1 478 NULL 2 -479 val_479 1 +478 val_478 2 479 NULL 1 -480 val_480 3 +479 val_479 1 480 NULL 3 -481 val_481 1 +480 val_480 3 481 NULL 1 -482 val_482 1 +481 val_481 1 482 NULL 1 -483 val_483 1 +482 val_482 1 483 NULL 1 -484 val_484 1 +483 val_483 1 484 NULL 1 -485 val_485 1 +484 val_484 1 485 NULL 1 -487 val_487 1 +485 val_485 1 487 NULL 1 -489 val_489 4 +487 val_487 1 489 NULL 4 -490 val_490 1 +489 val_489 4 490 NULL 1 -491 val_491 1 +490 val_490 1 491 NULL 1 -492 val_492 2 +491 val_491 1 492 NULL 2 -493 val_493 1 +492 val_492 2 493 NULL 1 -494 val_494 1 +493 val_493 1 494 NULL 1 -495 val_495 1 +494 val_494 1 495 NULL 1 -496 val_496 1 +495 val_495 1 496 NULL 1 -497 val_497 1 +496 val_496 1 497 NULL 1 -498 val_498 3 +497 val_497 1 498 NULL 3 -5 val_5 3 +498 val_498 3 5 NULL 3 -51 val_51 2 +5 val_5 3 51 NULL 2 -53 val_53 1 +51 val_51 2 53 NULL 1 -54 val_54 1 +53 val_53 1 54 NULL 1 -57 val_57 1 +54 val_54 1 57 NULL 1 -58 val_58 2 +57 val_57 1 58 NULL 2 -64 val_64 1 +58 val_58 2 64 NULL 1 -65 val_65 1 +64 val_64 1 65 NULL 1 -66 val_66 1 +65 val_65 1 66 NULL 1 -67 val_67 2 +66 val_66 1 67 NULL 2 -69 val_69 1 +67 val_67 2 69 NULL 1 -70 val_70 3 +69 val_69 1 70 NULL 3 -72 val_72 2 +70 val_70 3 72 NULL 2 -74 val_74 1 +72 val_72 2 74 NULL 1 -76 val_76 2 +74 val_74 1 76 NULL 2 -77 val_77 1 +76 val_76 2 77 NULL 1 -78 val_78 1 +77 val_77 1 78 NULL 1 -8 val_8 1 +78 val_78 1 8 NULL 1 -80 val_80 1 +8 val_8 1 80 NULL 1 -82 val_82 1 +80 val_80 1 82 NULL 1 -83 val_83 2 +82 val_82 1 83 NULL 2 -84 val_84 2 +83 val_83 2 84 NULL 2 -85 val_85 1 +84 val_84 2 85 NULL 1 -86 val_86 1 +85 val_85 1 86 NULL 1 -87 val_87 1 +86 val_86 1 87 NULL 1 -9 val_9 1 +87 val_87 1 9 NULL 1 -90 val_90 3 +9 val_9 1 90 NULL 3 -92 val_92 1 +90 val_90 3 92 NULL 1 -95 val_95 2 +92 val_92 1 95 NULL 2 -96 val_96 1 +95 val_95 2 96 NULL 1 -97 val_97 2 +96 val_96 1 97 NULL 2 -98 val_98 2 +97 val_97 2 98 NULL 2 +98 val_98 2 NULL NULL 500 PREHOOK: query: INSERT OVERWRITE TABLE test_table_out PARTITION (part = '1') SELECT key, value, count(1) FROM src GROUP BY key, value WITH ROLLUP @@ -807,7 +810,7 @@ Table: test_table_out #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 619 rawDataSize 6309 totalSize 6928 @@ -818,9 +821,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_2 PARTITION (part = '1') @@ -845,624 +848,624 @@ POSTHOOK: query: SELECT key, value, GROUPING__ID, count(1) FROM src GROUP BY ROL POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### -0 val_0 0 3 0 NULL 1 3 -10 val_10 0 1 +0 val_0 0 3 10 NULL 1 1 -100 val_100 0 2 +10 val_10 0 1 100 NULL 1 2 -103 val_103 0 2 +100 val_100 0 2 103 NULL 1 2 -104 val_104 0 2 +103 val_103 0 2 104 NULL 1 2 -105 val_105 0 1 +104 val_104 0 2 105 NULL 1 1 -11 val_11 0 1 +105 val_105 0 1 11 NULL 1 1 -111 val_111 0 1 +11 val_11 0 1 111 NULL 1 1 -113 val_113 0 2 +111 val_111 0 1 113 NULL 1 2 -114 val_114 0 1 +113 val_113 0 2 114 NULL 1 1 -116 val_116 0 1 +114 val_114 0 1 116 NULL 1 1 -118 val_118 0 2 +116 val_116 0 1 118 NULL 1 2 -119 val_119 0 3 +118 val_118 0 2 119 NULL 1 3 -12 val_12 0 2 +119 val_119 0 3 12 NULL 1 2 -120 val_120 0 2 +12 val_12 0 2 120 NULL 1 2 -125 val_125 0 2 +120 val_120 0 2 125 NULL 1 2 -126 val_126 0 1 +125 val_125 0 2 126 NULL 1 1 -128 val_128 0 3 +126 val_126 0 1 128 NULL 1 3 -129 val_129 0 2 +128 val_128 0 3 129 NULL 1 2 -131 val_131 0 1 +129 val_129 0 2 131 NULL 1 1 -133 val_133 0 1 +131 val_131 0 1 133 NULL 1 1 -134 val_134 0 2 +133 val_133 0 1 134 NULL 1 2 -136 val_136 0 1 +134 val_134 0 2 136 NULL 1 1 -137 val_137 0 2 +136 val_136 0 1 137 NULL 1 2 -138 val_138 0 4 +137 val_137 0 2 138 NULL 1 4 -143 val_143 0 1 +138 val_138 0 4 143 NULL 1 1 -145 val_145 0 1 +143 val_143 0 1 145 NULL 1 1 -146 val_146 0 2 +145 val_145 0 1 146 NULL 1 2 -149 val_149 0 2 +146 val_146 0 2 149 NULL 1 2 -15 val_15 0 2 +149 val_149 0 2 15 NULL 1 2 -150 val_150 0 1 +15 val_15 0 2 150 NULL 1 1 -152 val_152 0 2 +150 val_150 0 1 152 NULL 1 2 -153 val_153 0 1 +152 val_152 0 2 153 NULL 1 1 -155 val_155 0 1 +153 val_153 0 1 155 NULL 1 1 -156 val_156 0 1 +155 val_155 0 1 156 NULL 1 1 -157 val_157 0 1 +156 val_156 0 1 157 NULL 1 1 -158 val_158 0 1 +157 val_157 0 1 158 NULL 1 1 -160 val_160 0 1 +158 val_158 0 1 160 NULL 1 1 -162 val_162 0 1 +160 val_160 0 1 162 NULL 1 1 -163 val_163 0 1 +162 val_162 0 1 163 NULL 1 1 -164 val_164 0 2 +163 val_163 0 1 164 NULL 1 2 -165 val_165 0 2 +164 val_164 0 2 165 NULL 1 2 -166 val_166 0 1 +165 val_165 0 2 166 NULL 1 1 -167 val_167 0 3 +166 val_166 0 1 167 NULL 1 3 -168 val_168 0 1 +167 val_167 0 3 168 NULL 1 1 -169 val_169 0 4 +168 val_168 0 1 169 NULL 1 4 -17 val_17 0 1 +169 val_169 0 4 17 NULL 1 1 -170 val_170 0 1 +17 val_17 0 1 170 NULL 1 1 -172 val_172 0 2 +170 val_170 0 1 172 NULL 1 2 -174 val_174 0 2 +172 val_172 0 2 174 NULL 1 2 -175 val_175 0 2 +174 val_174 0 2 175 NULL 1 2 -176 val_176 0 2 +175 val_175 0 2 176 NULL 1 2 -177 val_177 0 1 +176 val_176 0 2 177 NULL 1 1 -178 val_178 0 1 +177 val_177 0 1 178 NULL 1 1 -179 val_179 0 2 +178 val_178 0 1 179 NULL 1 2 -18 val_18 0 2 +179 val_179 0 2 18 NULL 1 2 -180 val_180 0 1 +18 val_18 0 2 180 NULL 1 1 -181 val_181 0 1 +180 val_180 0 1 181 NULL 1 1 -183 val_183 0 1 +181 val_181 0 1 183 NULL 1 1 -186 val_186 0 1 +183 val_183 0 1 186 NULL 1 1 -187 val_187 0 3 +186 val_186 0 1 187 NULL 1 3 -189 val_189 0 1 +187 val_187 0 3 189 NULL 1 1 -19 val_19 0 1 +189 val_189 0 1 19 NULL 1 1 -190 val_190 0 1 +19 val_19 0 1 190 NULL 1 1 -191 val_191 0 2 +190 val_190 0 1 191 NULL 1 2 -192 val_192 0 1 +191 val_191 0 2 192 NULL 1 1 -193 val_193 0 3 +192 val_192 0 1 193 NULL 1 3 -194 val_194 0 1 +193 val_193 0 3 194 NULL 1 1 -195 val_195 0 2 +194 val_194 0 1 195 NULL 1 2 -196 val_196 0 1 +195 val_195 0 2 196 NULL 1 1 -197 val_197 0 2 +196 val_196 0 1 197 NULL 1 2 -199 val_199 0 3 +197 val_197 0 2 199 NULL 1 3 -2 val_2 0 1 +199 val_199 0 3 2 NULL 1 1 -20 val_20 0 1 +2 val_2 0 1 20 NULL 1 1 -200 val_200 0 2 +20 val_20 0 1 200 NULL 1 2 -201 val_201 0 1 +200 val_200 0 2 201 NULL 1 1 -202 val_202 0 1 +201 val_201 0 1 202 NULL 1 1 -203 val_203 0 2 +202 val_202 0 1 203 NULL 1 2 -205 val_205 0 2 +203 val_203 0 2 205 NULL 1 2 -207 val_207 0 2 +205 val_205 0 2 207 NULL 1 2 -208 val_208 0 3 +207 val_207 0 2 208 NULL 1 3 -209 val_209 0 2 +208 val_208 0 3 209 NULL 1 2 -213 val_213 0 2 +209 val_209 0 2 213 NULL 1 2 -214 val_214 0 1 +213 val_213 0 2 214 NULL 1 1 -216 val_216 0 2 +214 val_214 0 1 216 NULL 1 2 -217 val_217 0 2 +216 val_216 0 2 217 NULL 1 2 -218 val_218 0 1 +217 val_217 0 2 218 NULL 1 1 -219 val_219 0 2 +218 val_218 0 1 219 NULL 1 2 -221 val_221 0 2 +219 val_219 0 2 221 NULL 1 2 -222 val_222 0 1 +221 val_221 0 2 222 NULL 1 1 -223 val_223 0 2 +222 val_222 0 1 223 NULL 1 2 -224 val_224 0 2 +223 val_223 0 2 224 NULL 1 2 -226 val_226 0 1 +224 val_224 0 2 226 NULL 1 1 -228 val_228 0 1 +226 val_226 0 1 228 NULL 1 1 -229 val_229 0 2 +228 val_228 0 1 229 NULL 1 2 -230 val_230 0 5 +229 val_229 0 2 230 NULL 1 5 -233 val_233 0 2 +230 val_230 0 5 233 NULL 1 2 -235 val_235 0 1 +233 val_233 0 2 235 NULL 1 1 -237 val_237 0 2 +235 val_235 0 1 237 NULL 1 2 -238 val_238 0 2 +237 val_237 0 2 238 NULL 1 2 -239 val_239 0 2 +238 val_238 0 2 239 NULL 1 2 -24 val_24 0 2 +239 val_239 0 2 24 NULL 1 2 -241 val_241 0 1 +24 val_24 0 2 241 NULL 1 1 -242 val_242 0 2 +241 val_241 0 1 242 NULL 1 2 -244 val_244 0 1 +242 val_242 0 2 244 NULL 1 1 -247 val_247 0 1 +244 val_244 0 1 247 NULL 1 1 -248 val_248 0 1 +247 val_247 0 1 248 NULL 1 1 -249 val_249 0 1 +248 val_248 0 1 249 NULL 1 1 -252 val_252 0 1 +249 val_249 0 1 252 NULL 1 1 -255 val_255 0 2 +252 val_252 0 1 255 NULL 1 2 -256 val_256 0 2 +255 val_255 0 2 256 NULL 1 2 -257 val_257 0 1 +256 val_256 0 2 257 NULL 1 1 -258 val_258 0 1 +257 val_257 0 1 258 NULL 1 1 -26 val_26 0 2 +258 val_258 0 1 26 NULL 1 2 -260 val_260 0 1 +26 val_26 0 2 260 NULL 1 1 -262 val_262 0 1 +260 val_260 0 1 262 NULL 1 1 -263 val_263 0 1 +262 val_262 0 1 263 NULL 1 1 -265 val_265 0 2 +263 val_263 0 1 265 NULL 1 2 -266 val_266 0 1 +265 val_265 0 2 266 NULL 1 1 -27 val_27 0 1 +266 val_266 0 1 27 NULL 1 1 -272 val_272 0 2 +27 val_27 0 1 272 NULL 1 2 -273 val_273 0 3 +272 val_272 0 2 273 NULL 1 3 -274 val_274 0 1 +273 val_273 0 3 274 NULL 1 1 -275 val_275 0 1 +274 val_274 0 1 275 NULL 1 1 -277 val_277 0 4 +275 val_275 0 1 277 NULL 1 4 -278 val_278 0 2 +277 val_277 0 4 278 NULL 1 2 -28 val_28 0 1 +278 val_278 0 2 28 NULL 1 1 -280 val_280 0 2 +28 val_28 0 1 280 NULL 1 2 -281 val_281 0 2 +280 val_280 0 2 281 NULL 1 2 -282 val_282 0 2 +281 val_281 0 2 282 NULL 1 2 -283 val_283 0 1 +282 val_282 0 2 283 NULL 1 1 -284 val_284 0 1 +283 val_283 0 1 284 NULL 1 1 -285 val_285 0 1 +284 val_284 0 1 285 NULL 1 1 -286 val_286 0 1 +285 val_285 0 1 286 NULL 1 1 -287 val_287 0 1 +286 val_286 0 1 287 NULL 1 1 -288 val_288 0 2 +287 val_287 0 1 288 NULL 1 2 -289 val_289 0 1 +288 val_288 0 2 289 NULL 1 1 -291 val_291 0 1 +289 val_289 0 1 291 NULL 1 1 -292 val_292 0 1 +291 val_291 0 1 292 NULL 1 1 -296 val_296 0 1 +292 val_292 0 1 296 NULL 1 1 -298 val_298 0 3 +296 val_296 0 1 298 NULL 1 3 -30 val_30 0 1 +298 val_298 0 3 30 NULL 1 1 -302 val_302 0 1 +30 val_30 0 1 302 NULL 1 1 -305 val_305 0 1 +302 val_302 0 1 305 NULL 1 1 -306 val_306 0 1 +305 val_305 0 1 306 NULL 1 1 -307 val_307 0 2 +306 val_306 0 1 307 NULL 1 2 -308 val_308 0 1 +307 val_307 0 2 308 NULL 1 1 -309 val_309 0 2 +308 val_308 0 1 309 NULL 1 2 -310 val_310 0 1 +309 val_309 0 2 310 NULL 1 1 -311 val_311 0 3 +310 val_310 0 1 311 NULL 1 3 -315 val_315 0 1 +311 val_311 0 3 315 NULL 1 1 -316 val_316 0 3 +315 val_315 0 1 316 NULL 1 3 -317 val_317 0 2 +316 val_316 0 3 317 NULL 1 2 -318 val_318 0 3 +317 val_317 0 2 318 NULL 1 3 -321 val_321 0 2 +318 val_318 0 3 321 NULL 1 2 -322 val_322 0 2 +321 val_321 0 2 322 NULL 1 2 -323 val_323 0 1 +322 val_322 0 2 323 NULL 1 1 -325 val_325 0 2 +323 val_323 0 1 325 NULL 1 2 -327 val_327 0 3 +325 val_325 0 2 327 NULL 1 3 -33 val_33 0 1 +327 val_327 0 3 33 NULL 1 1 -331 val_331 0 2 +33 val_33 0 1 331 NULL 1 2 -332 val_332 0 1 +331 val_331 0 2 332 NULL 1 1 -333 val_333 0 2 +332 val_332 0 1 333 NULL 1 2 -335 val_335 0 1 +333 val_333 0 2 335 NULL 1 1 -336 val_336 0 1 +335 val_335 0 1 336 NULL 1 1 -338 val_338 0 1 +336 val_336 0 1 338 NULL 1 1 -339 val_339 0 1 +338 val_338 0 1 339 NULL 1 1 -34 val_34 0 1 +339 val_339 0 1 34 NULL 1 1 -341 val_341 0 1 +34 val_34 0 1 341 NULL 1 1 -342 val_342 0 2 +341 val_341 0 1 342 NULL 1 2 -344 val_344 0 2 +342 val_342 0 2 344 NULL 1 2 -345 val_345 0 1 +344 val_344 0 2 345 NULL 1 1 -348 val_348 0 5 +345 val_345 0 1 348 NULL 1 5 -35 val_35 0 3 +348 val_348 0 5 35 NULL 1 3 -351 val_351 0 1 +35 val_35 0 3 351 NULL 1 1 -353 val_353 0 2 +351 val_351 0 1 353 NULL 1 2 -356 val_356 0 1 +353 val_353 0 2 356 NULL 1 1 -360 val_360 0 1 +356 val_356 0 1 360 NULL 1 1 -362 val_362 0 1 +360 val_360 0 1 362 NULL 1 1 -364 val_364 0 1 +362 val_362 0 1 364 NULL 1 1 -365 val_365 0 1 +364 val_364 0 1 365 NULL 1 1 -366 val_366 0 1 +365 val_365 0 1 366 NULL 1 1 -367 val_367 0 2 +366 val_366 0 1 367 NULL 1 2 -368 val_368 0 1 +367 val_367 0 2 368 NULL 1 1 -369 val_369 0 3 +368 val_368 0 1 369 NULL 1 3 -37 val_37 0 2 +369 val_369 0 3 37 NULL 1 2 -373 val_373 0 1 +37 val_37 0 2 373 NULL 1 1 -374 val_374 0 1 +373 val_373 0 1 374 NULL 1 1 -375 val_375 0 1 +374 val_374 0 1 375 NULL 1 1 -377 val_377 0 1 +375 val_375 0 1 377 NULL 1 1 -378 val_378 0 1 +377 val_377 0 1 378 NULL 1 1 -379 val_379 0 1 +378 val_378 0 1 379 NULL 1 1 -382 val_382 0 2 +379 val_379 0 1 382 NULL 1 2 -384 val_384 0 3 +382 val_382 0 2 384 NULL 1 3 -386 val_386 0 1 +384 val_384 0 3 386 NULL 1 1 -389 val_389 0 1 +386 val_386 0 1 389 NULL 1 1 -392 val_392 0 1 +389 val_389 0 1 392 NULL 1 1 -393 val_393 0 1 +392 val_392 0 1 393 NULL 1 1 -394 val_394 0 1 +393 val_393 0 1 394 NULL 1 1 -395 val_395 0 2 +394 val_394 0 1 395 NULL 1 2 -396 val_396 0 3 +395 val_395 0 2 396 NULL 1 3 -397 val_397 0 2 +396 val_396 0 3 397 NULL 1 2 -399 val_399 0 2 +397 val_397 0 2 399 NULL 1 2 -4 val_4 0 1 +399 val_399 0 2 4 NULL 1 1 -400 val_400 0 1 +4 val_4 0 1 400 NULL 1 1 -401 val_401 0 5 +400 val_400 0 1 401 NULL 1 5 -402 val_402 0 1 +401 val_401 0 5 402 NULL 1 1 -403 val_403 0 3 +402 val_402 0 1 403 NULL 1 3 -404 val_404 0 2 +403 val_403 0 3 404 NULL 1 2 -406 val_406 0 4 +404 val_404 0 2 406 NULL 1 4 -407 val_407 0 1 +406 val_406 0 4 407 NULL 1 1 -409 val_409 0 3 +407 val_407 0 1 409 NULL 1 3 -41 val_41 0 1 +409 val_409 0 3 41 NULL 1 1 -411 val_411 0 1 +41 val_41 0 1 411 NULL 1 1 -413 val_413 0 2 +411 val_411 0 1 413 NULL 1 2 -414 val_414 0 2 +413 val_413 0 2 414 NULL 1 2 -417 val_417 0 3 +414 val_414 0 2 417 NULL 1 3 -418 val_418 0 1 +417 val_417 0 3 418 NULL 1 1 -419 val_419 0 1 +418 val_418 0 1 419 NULL 1 1 -42 val_42 0 2 +419 val_419 0 1 42 NULL 1 2 -421 val_421 0 1 +42 val_42 0 2 421 NULL 1 1 -424 val_424 0 2 +421 val_421 0 1 424 NULL 1 2 -427 val_427 0 1 +424 val_424 0 2 427 NULL 1 1 -429 val_429 0 2 +427 val_427 0 1 429 NULL 1 2 -43 val_43 0 1 +429 val_429 0 2 43 NULL 1 1 -430 val_430 0 3 +43 val_43 0 1 430 NULL 1 3 -431 val_431 0 3 +430 val_430 0 3 431 NULL 1 3 -432 val_432 0 1 +431 val_431 0 3 432 NULL 1 1 -435 val_435 0 1 +432 val_432 0 1 435 NULL 1 1 -436 val_436 0 1 +435 val_435 0 1 436 NULL 1 1 -437 val_437 0 1 +436 val_436 0 1 437 NULL 1 1 -438 val_438 0 3 +437 val_437 0 1 438 NULL 1 3 -439 val_439 0 2 +438 val_438 0 3 439 NULL 1 2 -44 val_44 0 1 +439 val_439 0 2 44 NULL 1 1 -443 val_443 0 1 +44 val_44 0 1 443 NULL 1 1 -444 val_444 0 1 +443 val_443 0 1 444 NULL 1 1 -446 val_446 0 1 +444 val_444 0 1 446 NULL 1 1 -448 val_448 0 1 +446 val_446 0 1 448 NULL 1 1 -449 val_449 0 1 +448 val_448 0 1 449 NULL 1 1 -452 val_452 0 1 +449 val_449 0 1 452 NULL 1 1 -453 val_453 0 1 +452 val_452 0 1 453 NULL 1 1 -454 val_454 0 3 +453 val_453 0 1 454 NULL 1 3 -455 val_455 0 1 +454 val_454 0 3 455 NULL 1 1 -457 val_457 0 1 +455 val_455 0 1 457 NULL 1 1 -458 val_458 0 2 +457 val_457 0 1 458 NULL 1 2 -459 val_459 0 2 +458 val_458 0 2 459 NULL 1 2 -460 val_460 0 1 +459 val_459 0 2 460 NULL 1 1 -462 val_462 0 2 +460 val_460 0 1 462 NULL 1 2 -463 val_463 0 2 +462 val_462 0 2 463 NULL 1 2 -466 val_466 0 3 +463 val_463 0 2 466 NULL 1 3 -467 val_467 0 1 +466 val_466 0 3 467 NULL 1 1 -468 val_468 0 4 +467 val_467 0 1 468 NULL 1 4 -469 val_469 0 5 +468 val_468 0 4 469 NULL 1 5 -47 val_47 0 1 +469 val_469 0 5 47 NULL 1 1 -470 val_470 0 1 +47 val_47 0 1 470 NULL 1 1 -472 val_472 0 1 +470 val_470 0 1 472 NULL 1 1 -475 val_475 0 1 +472 val_472 0 1 475 NULL 1 1 -477 val_477 0 1 +475 val_475 0 1 477 NULL 1 1 -478 val_478 0 2 +477 val_477 0 1 478 NULL 1 2 -479 val_479 0 1 +478 val_478 0 2 479 NULL 1 1 -480 val_480 0 3 +479 val_479 0 1 480 NULL 1 3 -481 val_481 0 1 +480 val_480 0 3 481 NULL 1 1 -482 val_482 0 1 +481 val_481 0 1 482 NULL 1 1 -483 val_483 0 1 +482 val_482 0 1 483 NULL 1 1 -484 val_484 0 1 +483 val_483 0 1 484 NULL 1 1 -485 val_485 0 1 +484 val_484 0 1 485 NULL 1 1 -487 val_487 0 1 +485 val_485 0 1 487 NULL 1 1 -489 val_489 0 4 +487 val_487 0 1 489 NULL 1 4 -490 val_490 0 1 +489 val_489 0 4 490 NULL 1 1 -491 val_491 0 1 +490 val_490 0 1 491 NULL 1 1 -492 val_492 0 2 +491 val_491 0 1 492 NULL 1 2 -493 val_493 0 1 +492 val_492 0 2 493 NULL 1 1 -494 val_494 0 1 +493 val_493 0 1 494 NULL 1 1 -495 val_495 0 1 +494 val_494 0 1 495 NULL 1 1 -496 val_496 0 1 +495 val_495 0 1 496 NULL 1 1 -497 val_497 0 1 +496 val_496 0 1 497 NULL 1 1 -498 val_498 0 3 +497 val_497 0 1 498 NULL 1 3 -5 val_5 0 3 +498 val_498 0 3 5 NULL 1 3 -51 val_51 0 2 +5 val_5 0 3 51 NULL 1 2 -53 val_53 0 1 +51 val_51 0 2 53 NULL 1 1 -54 val_54 0 1 +53 val_53 0 1 54 NULL 1 1 -57 val_57 0 1 +54 val_54 0 1 57 NULL 1 1 -58 val_58 0 2 +57 val_57 0 1 58 NULL 1 2 -64 val_64 0 1 +58 val_58 0 2 64 NULL 1 1 -65 val_65 0 1 +64 val_64 0 1 65 NULL 1 1 -66 val_66 0 1 +65 val_65 0 1 66 NULL 1 1 -67 val_67 0 2 +66 val_66 0 1 67 NULL 1 2 -69 val_69 0 1 +67 val_67 0 2 69 NULL 1 1 -70 val_70 0 3 +69 val_69 0 1 70 NULL 1 3 -72 val_72 0 2 +70 val_70 0 3 72 NULL 1 2 -74 val_74 0 1 +72 val_72 0 2 74 NULL 1 1 -76 val_76 0 2 +74 val_74 0 1 76 NULL 1 2 -77 val_77 0 1 +76 val_76 0 2 77 NULL 1 1 -78 val_78 0 1 +77 val_77 0 1 78 NULL 1 1 -8 val_8 0 1 +78 val_78 0 1 8 NULL 1 1 -80 val_80 0 1 +8 val_8 0 1 80 NULL 1 1 -82 val_82 0 1 +80 val_80 0 1 82 NULL 1 1 -83 val_83 0 2 +82 val_82 0 1 83 NULL 1 2 -84 val_84 0 2 +83 val_83 0 2 84 NULL 1 2 -85 val_85 0 1 +84 val_84 0 2 85 NULL 1 1 -86 val_86 0 1 +85 val_85 0 1 86 NULL 1 1 -87 val_87 0 1 +86 val_86 0 1 87 NULL 1 1 -9 val_9 0 1 +87 val_87 0 1 9 NULL 1 1 -90 val_90 0 3 +9 val_9 0 1 90 NULL 1 3 -92 val_92 0 1 +90 val_90 0 3 92 NULL 1 1 -95 val_95 0 2 +92 val_92 0 1 95 NULL 1 2 -96 val_96 0 1 +95 val_95 0 2 96 NULL 1 1 -97 val_97 0 2 +96 val_96 0 1 97 NULL 1 2 -98 val_98 0 2 +97 val_97 0 2 98 NULL 1 2 +98 val_98 0 2 NULL NULL 3 500 PREHOOK: query: DESCRIBE FORMATTED test_table_out_2 PARTITION (part = '1') PREHOOK: type: DESCTABLE @@ -1487,7 +1490,7 @@ Table: test_table_out_2 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"grouping_key\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 619 rawDataSize 7547 totalSize 8166 @@ -1515,73 +1518,107 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@test_table_out@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), value (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1000 Data size: 194000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 1000 Data size: 194000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - keys: key (type: string), value (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 1000 Data size: 194000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 1000 Data size: 194000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 1000 Data size: 194000 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) - outputColumnNames: key, value, agg, part - Statistics: Num rows: 1000 Data size: 447000 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) + outputColumnNames: key, value, agg, part + Statistics: Num rows: 1000 Data size: 447000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') + keys: part (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') - keys: part (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -1595,7 +1632,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_out - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -1603,37 +1640,6 @@ STAGE PLANS: Column Types: string, string, string Table: default.test_table_out - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out PARTITION (part = '1') SELECT key, value, count(1) FROM src GROUP BY key, value WITH CUBE PREHOOK: type: QUERY @@ -1669,7 +1675,7 @@ Table: test_table_out #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 928 rawDataSize 9954 totalSize 10882 @@ -1680,9 +1686,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_2 PARTITION (part = '1') @@ -1722,7 +1728,7 @@ Table: test_table_out_2 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"grouping_key\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 928 rawDataSize 11810 totalSize 12738 @@ -1750,73 +1756,107 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@test_table_out@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string), value (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - keys: key (type: string), value (type: string), 0L (type: bigint) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3 Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - null sort order: zzz - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) - Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col3 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) - mode: mergepartial - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 500 Data size: 97000 Basic stats: COMPLETE Column stats: COMPLETE - pruneGroupingSetId: true - Select Operator - expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 181000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 181000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) - outputColumnNames: key, value, agg, part - Statistics: Num rows: 500 Data size: 223500 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), CAST( _col3 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 181000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 181000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), '1' (type: string) + outputColumnNames: key, value, agg, part + Statistics: Num rows: 500 Data size: 223500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') + keys: part (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(agg, 'hll') - keys: part (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -1830,7 +1870,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_out - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -1838,37 +1878,6 @@ STAGE PLANS: Column Types: string, string, string Table: default.test_table_out - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1405 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out PARTITION (part = '1') SELECT key, value, count(1) FROM src GROUP BY key, value GROUPING SETS (key, value) PREHOOK: type: QUERY @@ -1904,7 +1913,7 @@ Table: test_table_out #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 618 rawDataSize 6054 totalSize 6672 @@ -1915,9 +1924,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_2 PARTITION (part = '1') @@ -1957,7 +1966,7 @@ Table: test_table_out_2 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"agg\":\"true\",\"grouping_key\":\"true\",\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 618 rawDataSize 7290 totalSize 7908 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_list_bucket.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_list_bucket.q.out similarity index 99% rename from ql/src/test/results/clientpositive/infer_bucket_sort_list_bucket.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_list_bucket.q.out index c701de9347..44d10baac4 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_list_bucket.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_list_bucket.q.out @@ -45,7 +45,7 @@ Table: list_bucketing_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 2 + numFiles 3 numRows 309 rawDataSize 1482 totalSize 1791 @@ -113,7 +113,7 @@ Table: list_bucketing_table2 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1482 totalSize 1791 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_map_operators.q.out similarity index 53% rename from ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_map_operators.q.out index e2234c302b..b6c7644a59 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_map_operators.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_map_operators.q.out @@ -58,88 +58,88 @@ POSTHOOK: Input: default@test_table1_n14 POSTHOOK: Output: default@test_table_out_n0@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: test_table1_n14 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: test_table1_n14 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 316 Data size: 85636 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 316 Data size: 85636 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '1' (type: string) + outputColumnNames: key, value, part + Statistics: Num rows: 316 Data size: 112496 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: part (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: count() - keys: key (type: string) - mode: final - outputColumnNames: _col0, _col1 - Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 316 Data size: 85636 Basic stats: COMPLETE Column stats: COMPLETE + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 316 Data size: 85636 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '1' (type: string) - outputColumnNames: key, value, part - Statistics: Num rows: 316 Data size: 112496 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: part (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -153,7 +153,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_out_n0 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -161,36 +161,6 @@ STAGE PLANS: Column Types: string, string Table: default.test_table_out_n0 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1') SELECT key, count(*) FROM test_table1_n14 GROUP BY key PREHOOK: type: QUERY @@ -224,7 +194,7 @@ Table: test_table_out_n0 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1482 totalSize 1791 @@ -265,96 +235,135 @@ POSTHOOK: Input: default@test_table1_n14 POSTHOOK: Output: default@test_table_out_n0@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: test_table1_n14 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: key - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - keys: key (type: string) - mode: final + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: test_table1_n14 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: key + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + mode: final + outputColumnNames: _col0, _col1 + Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(_col1) is not null (type: boolean) + Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: bigint), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 316 Data size: 32548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: double) + Statistics: Num rows: 316 Data size: 32548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(value) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(value) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToDouble(value) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: double) + 1 _col0 (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(_col1) is not null (type: boolean) - Statistics: Num rows: 316 Data size: 30020 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 135500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 135500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out_n0 Select Operator - expressions: _col0 (type: string), _col1 (type: bigint), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 316 Data size: 32548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: double) - Statistics: Num rows: 316 Data size: 32548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: bigint) - TableScan - alias: src - filterExpr: UDFToDouble(value) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(value) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToDouble(value) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: double) - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: double) - 1 _col0 (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 135500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 135500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '1' (type: string) - outputColumnNames: key, value, part - Statistics: Num rows: 500 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + expressions: _col0 (type: string), _col1 (type: string), '1' (type: string) + outputColumnNames: key, value, part + Statistics: Num rows: 500 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: part (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: part (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -368,7 +377,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_out_n0 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -376,37 +385,6 @@ STAGE PLANS: Column Types: string, string Table: default.test_table_out_n0 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1') SELECT a.key, a.value FROM ( SELECT key, cast(count(*) AS STRING) AS value FROM test_table1_n14 GROUP BY key @@ -452,7 +430,7 @@ Table: test_table_out_n0 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 0 numRows 0 rawDataSize 0 totalSize 0 @@ -463,9 +441,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1') @@ -482,90 +460,98 @@ POSTHOOK: Input: default@test_table2_n13 POSTHOOK: Output: default@test_table_out_n0@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-8 depends on stages: Stage-1 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Sorted Merge Bucket Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col0, _col6 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col6 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 791 Data size: 208033 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: '1' (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: '1' (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: '1' (type: string) + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: '1' (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: _col0 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + expressions: _col1 (type: struct), _col2 (type: struct), '1' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: '1' (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: '1' (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: '1' (type: string) - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: struct), _col2 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: '1' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), '1' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -587,36 +573,6 @@ STAGE PLANS: Column Types: string, string Table: default.test_table_out_n0 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1') SELECT /*+ MAPJOIN(a) */ a.key, b.value FROM test_table1_n14 a JOIN test_table2_n13 b ON a.key = b.key PREHOOK: type: QUERY @@ -652,7 +608,7 @@ Table: test_table_out_n0 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 1028 rawDataSize 10968 totalSize 11996 @@ -684,79 +640,122 @@ POSTHOOK: Input: default@test_table2_n13 POSTHOOK: Output: default@test_table_out_n0@part=1 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Sorted Merge Bucket Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 key (type: string) - outputColumnNames: _col6 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: count() - keys: _col6 (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Dummy Store + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col6 + Statistics: Num rows: 791 Data size: 71981 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col6 (type: string) + minReductionHashAggr: 0.6118837 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: llap + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), CAST( _col1 AS STRING) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.test_table_out_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 307 Data size: 84425 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 84425 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_out_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 307 Data size: 110520 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: '1' (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: '1' (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: '1' (type: string) + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) keys: '1' (type: string) - minReductionHashAggr: 0.99 - mode: hash + mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), '1' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 965 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -778,37 +777,6 @@ STAGE PLANS: Column Types: string, string Table: default.test_table_out_n0 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: '1' (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: '1' (type: string) - Statistics: Num rows: 275 Data size: 23925 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: '1' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 137 Data size: 11919 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), '1' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 137 Data size: 11919 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 137 Data size: 11919 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE test_table_out_n0 PARTITION (part = '1') SELECT /*+ MAPJOIN(a) */ b.value, count(*) FROM test_table1_n14 a JOIN test_table2_n13 b ON a.key = b.key GROUP BY b.value @@ -846,7 +814,7 @@ Table: test_table_out_n0 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 2728 totalSize 3037 @@ -857,8 +825,8 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_merge.q.out similarity index 96% rename from ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_merge.q.out index 079039322d..38227454f1 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_merge.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_merge.q.out @@ -39,7 +39,7 @@ Table: test_table_n16 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 1028 rawDataSize 10968 totalSize 11996 @@ -99,8 +99,8 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 2 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_multi_insert.q.out similarity index 94% rename from ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_multi_insert.q.out index a10c06eba6..bc8f153166 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_multi_insert.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_multi_insert.q.out @@ -137,7 +137,7 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1482 totalSize 1791 @@ -148,9 +148,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -174,7 +174,7 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 2718 totalSize 3027 @@ -185,9 +185,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: FROM src @@ -229,7 +229,7 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1482 totalSize 1791 @@ -240,9 +240,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -321,7 +321,7 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 1482 totalSize 1791 @@ -332,9 +332,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (part = '2') @@ -358,7 +358,7 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 1 + numFiles 2 numRows 309 rawDataSize 2690 totalSize 2999 @@ -369,8 +369,8 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 diff --git ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out new file mode 100644 index 0000000000..6deaafcb57 --- /dev/null +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out @@ -0,0 +1,439 @@ +PREHOOK: query: CREATE TABLE test_table_n0 (key INT, value STRING) PARTITIONED BY (ds STRING, hr STRING) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_table_n0 +POSTHOOK: query: CREATE TABLE test_table_n0 (key INT, value STRING) PARTITIONED BY (ds STRING, hr STRING) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@test_table_n0 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) +SELECT key2, value, cast(hr as int) FROM +(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 2) as hr +FROM srcpart +WHERE ds = '2008-04-08') a +DISTRIBUTE BY key2 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@test_table_n0@ds=2008-04-08 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) +SELECT key2, value, cast(hr as int) FROM +(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 2) as hr +FROM srcpart +WHERE ds = '2008-04-08') a +DISTRIBUTE BY key2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: if(((key % 3) < 2), 0, 1) (type: int), value (type: string), UDFToInteger((key % 2)) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 99000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1000 Data size: 99000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int) + auto parallelism: false + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), CAST( VALUE._col2 AS STRING) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, hr + Statistics: Num rows: 1000 Data size: 373000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: '2008-04-08' (type: string), hr (type: string) + minReductionHashAggr: 0.684 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 316 Data size: 360872 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: '2008-04-08' (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: '2008-04-08' (type: string), _col1 (type: string) + Statistics: Num rows: 316 Data size: 360872 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col2 (type: string) + null sort order: a + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: int), _col1 (type: string) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: '2008-04-08' (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), '2008-04-08' (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 316 Data size: 365928 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 4 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 279000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.test_table_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct test_table_n0 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008-04-08 + hr + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.test_table_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct test_table_n0 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.test_table_n0 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.test_table_n0 + Is Table Level Stats: false + +PREHOOK: query: INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) +SELECT key2, value, cast(hr as int) FROM +(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 3 % 2) as hr +FROM srcpart +WHERE ds = '2008-04-08') a +DISTRIBUTE BY key2 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@test_table_n0@ds=2008-04-08 +POSTHOOK: query: INSERT OVERWRITE TABLE test_table_n0 PARTITION (ds = '2008-04-08', hr) +SELECT key2, value, cast(hr as int) FROM +(SELECT if ((key % 3) < 2, 0, 1) as key2, value, (key % 3 % 2) as hr +FROM srcpart +WHERE ds = '2008-04-08') a +DISTRIBUTE BY key2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@test_table_n0@ds=2008-04-08/hr=0 +POSTHOOK: Output: default@test_table_n0@ds=2008-04-08/hr=1 +POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=0).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=0).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=1).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: test_table_n0 PARTITION(ds=2008-04-08,hr=1).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='0') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@test_table_n0 +POSTHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='0') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@test_table_n0 +# col_name data_type comment +key int +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 0] +Database: default +Table: test_table_n0 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 1 + numRows 670 + rawDataSize 5888 + totalSize 6558 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 +PREHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='1') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@test_table_n0 +POSTHOOK: query: DESCRIBE FORMATTED test_table_n0 PARTITION (ds='2008-04-08', hr='1') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@test_table_n0 +# col_name data_type comment +key int +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 1] +Database: default +Table: test_table_n0 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 1 + numRows 330 + rawDataSize 2924 + totalSize 3254 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Storage Desc Params: + serialization.format 1 diff --git ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out ql/src/test/results/clientpositive/llap/infer_bucket_sort_reducers_power_two.q.out similarity index 94% rename from ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out rename to ql/src/test/results/clientpositive/llap/infer_bucket_sort_reducers_power_two.q.out index aa0d49a6bb..a72dc51bce 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_reducers_power_two.q.out +++ ql/src/test/results/clientpositive/llap/infer_bucket_sort_reducers_power_two.q.out @@ -39,7 +39,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 4 + numFiles 10 numRows 309 rawDataSize 1482 totalSize 1791 @@ -50,9 +50,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 4 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_n14 PARTITION (part = '1') @@ -88,7 +88,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 4 + numFiles 53 numRows 1028 rawDataSize 10968 totalSize 11996 @@ -99,9 +99,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 4 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_n14 PARTITION (part = '1') @@ -137,7 +137,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 4 + numFiles 71 numRows 1028 rawDataSize 10968 totalSize 11996 @@ -148,9 +148,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 4 -Bucket Columns: [key, value] -Sort Columns: [Order(col:key, order:1), Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_n14 PARTITION (part = '1') @@ -186,7 +186,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 13 + numFiles 88 numRows 2654 rawDataSize 28466 totalSize 31120 @@ -235,7 +235,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 16 + numFiles 74 numRows 2654 rawDataSize 28466 totalSize 31120 @@ -246,9 +246,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe InputFormat: org.apache.hadoop.mapred.TextInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Compressed: No -Num Buckets: 16 -Bucket Columns: [value] -Sort Columns: [Order(col:value, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: INSERT OVERWRITE TABLE test_table_n14 PARTITION (part = '1') @@ -284,7 +284,7 @@ Table: test_table_n14 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 1 + numFiles 2 numRows 5 rawDataSize 19 totalSize 24 diff --git ql/src/test/results/clientpositive/infer_const_type.q.out ql/src/test/results/clientpositive/llap/infer_const_type.q.out similarity index 58% rename from ql/src/test/results/clientpositive/infer_const_type.q.out rename to ql/src/test/results/clientpositive/llap/infer_const_type.q.out index b636090081..4eeeb7d1b9 100644 --- ql/src/test/results/clientpositive/infer_const_type.q.out +++ ql/src/test/results/clientpositive/llap/infer_const_type.q.out @@ -53,38 +53,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@infertypes #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: infertypes - filterExpr: ((ti = 127Y) and (si = 32767S) and (i = 12345) and (bi = -12345L) and (fl = 906.0) and (db = -307.0D) and (UDFToDouble(str) = 1234.0D)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((ti = 127Y) and (si = 32767S) and (i = 12345) and (bi = -12345L) and (fl = 906.0) and (db = -307.0D) and (UDFToDouble(str) = 1234.0D)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: 127Y (type: tinyint), 32767S (type: smallint), 12345 (type: int), -12345L (type: bigint), 906.0 (type: float), -307.0D (type: double), str (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: infertypes + filterExpr: ((ti = 127Y) and (si = 32767S) and (i = 12345) and (bi = -12345L) and (fl = 906.0) and (db = -307.0D) and (UDFToDouble(str) = 1234.0D)) (type: boolean) + Filter Operator + predicate: ((ti = 127Y) and (si = 32767S) and (i = 12345) and (bi = -12345L) and (fl = 906.0) and (db = -307.0D) and (UDFToDouble(str) = 1234.0D)) (type: boolean) + Select Operator + expressions: 127Y (type: tinyint), 32767S (type: smallint), 12345 (type: int), -12345L (type: bigint), 906.0 (type: float), -307.0D (type: double), str (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink WARNING: Comparing a bigint and a string may result in a loss of precision. PREHOOK: query: SELECT * FROM infertypes WHERE @@ -177,38 +161,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@infertypes #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: infertypes - filterExpr: ((ti = 127Y) or (si = 327S) or (i = -100)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((ti = 127Y) or (si = 327S) or (i = -100)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: ti (type: tinyint), si (type: smallint), i (type: int), bi (type: bigint), fl (type: float), db (type: double), str (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: infertypes + filterExpr: ((ti = 127Y) or (si = 327S) or (i = -100)) (type: boolean) + Filter Operator + predicate: ((ti = 127Y) or (si = 327S) or (i = -100)) (type: boolean) + Select Operator + expressions: ti (type: tinyint), si (type: smallint), i (type: int), bi (type: bigint), fl (type: float), db (type: double), str (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink PREHOOK: query: SELECT * FROM infertypes WHERE ti = '127.0' OR @@ -240,38 +208,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@infertypes #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: infertypes - filterExpr: ((ti < 127Y) and (i > 100) and (UDFToDouble(str) = 1.57D)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((ti < 127Y) and (i > 100) and (UDFToDouble(str) = 1.57D)) (type: boolean) - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: ti (type: tinyint), si (type: smallint), i (type: int), bi (type: bigint), fl (type: float), db (type: double), str (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 216 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: infertypes + filterExpr: ((ti < 127Y) and (i > 100) and (UDFToDouble(str) = 1.57D)) (type: boolean) + Filter Operator + predicate: ((ti < 127Y) and (i > 100) and (UDFToDouble(str) = 1.57D)) (type: boolean) + Select Operator + expressions: ti (type: tinyint), si (type: smallint), i (type: int), bi (type: bigint), fl (type: float), db (type: double), str (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink PREHOOK: query: SELECT * FROM infertypes WHERE ti < '127.0' AND diff --git ql/src/test/results/clientpositive/llap/infer_join_preds.q.out ql/src/test/results/clientpositive/llap/infer_join_preds.q.out new file mode 100644 index 0000000000..2d8653cada --- /dev/null +++ ql/src/test/results/clientpositive/llap/infer_join_preds.q.out @@ -0,0 +1,1433 @@ +PREHOOK: query: explain +select * from src a join src1 b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from src a join src1 b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from src a join src1 b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from src a join src1 b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 val_128 128 +128 val_128 128 +128 val_128 128 +146 val_146 146 val_146 +146 val_146 146 val_146 +150 val_150 150 val_150 +213 val_213 213 val_213 +213 val_213 213 val_213 +224 val_224 224 +224 val_224 224 +238 val_238 238 val_238 +238 val_238 238 val_238 +255 val_255 255 val_255 +255 val_255 255 val_255 +273 val_273 273 val_273 +273 val_273 273 val_273 +273 val_273 273 val_273 +278 val_278 278 val_278 +278 val_278 278 val_278 +311 val_311 311 val_311 +311 val_311 311 val_311 +311 val_311 311 val_311 +369 val_369 369 +369 val_369 369 +369 val_369 369 +401 val_401 401 val_401 +401 val_401 401 val_401 +401 val_401 401 val_401 +401 val_401 401 val_401 +401 val_401 401 val_401 +406 val_406 406 val_406 +406 val_406 406 val_406 +406 val_406 406 val_406 +406 val_406 406 val_406 +66 val_66 66 val_66 +98 val_98 98 val_98 +98 val_98 98 val_98 +PREHOOK: query: explain +select * from +(select * from src where 1 = 0)a +join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +(select * from src where 1 = 0)a +join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 353 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 353 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from +(select * from src where 1 = 0)a +join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from +(select * from src where 1 = 0)a +join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +PREHOOK: query: explain +select * from +(select * from src where 1 = 0)a +left outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +(select * from src where 1 = 0)a +left outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from +(select * from src where 1 = 0)a +left outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from +(select * from src where 1 = 0)a +left outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +PREHOOK: query: explain +select * from +(select * from src where 1 = 0)a +right outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +(select * from src where 1 = 0)a +right outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 26 Data size: 4906 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 26 Data size: 4906 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from +(select * from src where 1 = 0)a +right outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from +(select * from src where 1 = 0)a +right outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +NULL NULL +NULL NULL +NULL NULL +NULL NULL +NULL NULL val_165 +NULL NULL val_193 +NULL NULL val_265 +NULL NULL val_27 +NULL NULL val_409 +NULL NULL val_484 +NULL NULL 128 +NULL NULL 146 val_146 +NULL NULL 150 val_150 +NULL NULL 213 val_213 +NULL NULL 224 +NULL NULL 238 val_238 +NULL NULL 255 val_255 +NULL NULL 273 val_273 +NULL NULL 278 val_278 +NULL NULL 311 val_311 +NULL NULL 369 +NULL NULL 401 val_401 +NULL NULL 406 val_406 +NULL NULL 66 val_66 +NULL NULL 98 val_98 +PREHOOK: query: explain +select * from +(select * from src where 1 = 0)a +full outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +(select * from src where 1 = 0)a +full outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 27 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 27 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from +(select * from src where 1 = 0)a +full outer join +(select * from src1)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from +(select * from src where 1 = 0)a +full outer join +(select * from src1)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +NULL NULL +NULL NULL +NULL NULL +NULL NULL +NULL NULL val_165 +NULL NULL val_193 +NULL NULL val_265 +NULL NULL val_27 +NULL NULL val_409 +NULL NULL val_484 +NULL NULL 128 +NULL NULL 146 val_146 +NULL NULL 150 val_150 +NULL NULL 213 val_213 +NULL NULL 224 +NULL NULL 238 val_238 +NULL NULL 255 val_255 +NULL NULL 273 val_273 +NULL NULL 278 val_278 +NULL NULL 311 val_311 +NULL NULL 369 +NULL NULL 401 val_401 +NULL NULL 406 val_406 +NULL NULL 66 val_66 +NULL NULL 98 val_98 +PREHOOK: query: explain +select * from +(select * from src)a +right outer join +(select * from src1 where 1 = 0)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +(select * from src)a +right outer join +(select * from src1 where 1 = 0)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from +(select * from src)a +right outer join +(select * from src1 where 1 = 0)b on a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: select * from +(select * from src)a +right outer join +(select * from src1 where 1 = 0)b on a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +PREHOOK: query: explain select * from src join src1 on src.key = src1.key and src.value = src1.value + where 4 between src.key and src.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from src join src1 on src.key = src1.key and src.value = src1.value + where 4 between src.key and src.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: 4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: 4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) and key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (4.0D BETWEEN UDFToDouble(key) AND UDFToDouble(value) and key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col1 (type: string) + 1 _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: CREATE TABLE `table1_n8`( + `idp_warehouse_id` bigint, + `idp_audit_id` bigint, + `idp_effective_date` date, + `idp_end_date` date, + `idp_delete_date` date, + `pruid` varchar(32), + `prid` bigint, + `prtimesheetid` bigint, + `prassignmentid` bigint, + `prchargecodeid` bigint, + `prtypecodeid` bigint, + `prsequence` bigint, + `prmodby` varchar(96), + `prmodtime` timestamp, + `prrmexported` bigint, + `prrmckdel` bigint, + `slice_status` int, + `role_id` bigint, + `user_lov1` varchar(30), + `user_lov2` varchar(30), + `incident_id` bigint, + `incident_investment_id` bigint, + `odf_ss_actuals` bigint, + `practsum` decimal(38,20)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@table1_n8 +POSTHOOK: query: CREATE TABLE `table1_n8`( + `idp_warehouse_id` bigint, + `idp_audit_id` bigint, + `idp_effective_date` date, + `idp_end_date` date, + `idp_delete_date` date, + `pruid` varchar(32), + `prid` bigint, + `prtimesheetid` bigint, + `prassignmentid` bigint, + `prchargecodeid` bigint, + `prtypecodeid` bigint, + `prsequence` bigint, + `prmodby` varchar(96), + `prmodtime` timestamp, + `prrmexported` bigint, + `prrmckdel` bigint, + `slice_status` int, + `role_id` bigint, + `user_lov1` varchar(30), + `user_lov2` varchar(30), + `incident_id` bigint, + `incident_investment_id` bigint, + `odf_ss_actuals` bigint, + `practsum` decimal(38,20)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@table1_n8 +PREHOOK: query: CREATE TABLE `table2_n4`( + `idp_warehouse_id` bigint, + `idp_audit_id` bigint, + `idp_effective_date` date, + `idp_end_date` date, + `idp_delete_date` date, + `pruid` varchar(32), + `prid` bigint, + `prtimesheetid` bigint, + `prassignmentid` bigint, + `prchargecodeid` bigint, + `prtypecodeid` bigint, + `prsequence` bigint, + `prmodby` varchar(96), + `prmodtime` timestamp, + `prrmexported` bigint, + `prrmckdel` bigint, + `slice_status` int, + `role_id` bigint, + `user_lov1` varchar(30), + `user_lov2` varchar(30), + `incident_id` bigint, + `incident_investment_id` bigint, + `odf_ss_actuals` bigint, + `practsum` decimal(38,20)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@table2_n4 +POSTHOOK: query: CREATE TABLE `table2_n4`( + `idp_warehouse_id` bigint, + `idp_audit_id` bigint, + `idp_effective_date` date, + `idp_end_date` date, + `idp_delete_date` date, + `pruid` varchar(32), + `prid` bigint, + `prtimesheetid` bigint, + `prassignmentid` bigint, + `prchargecodeid` bigint, + `prtypecodeid` bigint, + `prsequence` bigint, + `prmodby` varchar(96), + `prmodtime` timestamp, + `prrmexported` bigint, + `prrmckdel` bigint, + `slice_status` int, + `role_id` bigint, + `user_lov1` varchar(30), + `user_lov2` varchar(30), + `incident_id` bigint, + `incident_investment_id` bigint, + `odf_ss_actuals` bigint, + `practsum` decimal(38,20)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@table2_n4 +PREHOOK: query: explain SELECT s.idp_warehouse_id AS source_warehouse_id + FROM table1_n8 s + JOIN + + table2_n4 d + ON ( + s.prid = d.prid ) + JOIN + table2_n4 e + ON + s.prid = e.prid + WHERE + concat( + CASE + WHEN s.prid IS NULL THEN 1 + ELSE s.prid + END,',', + CASE + WHEN s.prtimesheetid IS NULL THEN 1 + ELSE s.prtimesheetid + END,',', + CASE + WHEN s.prassignmentid IS NULL THEN 1 + ELSE s.prassignmentid + END,',', + CASE + WHEN s.prchargecodeid IS NULL THEN 1 + ELSE s.prchargecodeid + END,',', + CASE + WHEN (s.prtypecodeid) IS NULL THEN '' + ELSE s.prtypecodeid + END,',', + CASE + WHEN s.practsum IS NULL THEN 1 + ELSE s.practsum + END,',', + CASE + WHEN s.prsequence IS NULL THEN 1 + ELSE s.prsequence + END,',', + CASE + WHEN length(s.prmodby) IS NULL THEN '' + ELSE s.prmodby + END,',', + CASE + WHEN s.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) + ELSE s.prmodtime + END,',', + CASE + WHEN s.prrmexported IS NULL THEN 1 + ELSE s.prrmexported + END,',', + CASE + WHEN s.prrmckdel IS NULL THEN 1 + ELSE s.prrmckdel + END,',', + CASE + WHEN s.slice_status IS NULL THEN 1 + ELSE s.slice_status + END,',', + CASE + WHEN s.role_id IS NULL THEN 1 + ELSE s.role_id + END,',', + CASE + WHEN length(s.user_lov1) IS NULL THEN '' + ELSE s.user_lov1 + END,',', + CASE + WHEN length(s.user_lov2) IS NULL THEN '' + ELSE s.user_lov2 + END,',', + CASE + WHEN s.incident_id IS NULL THEN 1 + ELSE s.incident_id + END,',', + CASE + WHEN s.incident_investment_id IS NULL THEN 1 + ELSE s.incident_investment_id + END,',', + CASE + WHEN s.odf_ss_actuals IS NULL THEN 1 + ELSE s.odf_ss_actuals + END ) != concat( + CASE + WHEN length(d.pruid) IS NULL THEN '' + ELSE d.pruid + END,',', + CASE + WHEN d.prid IS NULL THEN 1 + ELSE d.prid + END,',', + CASE + WHEN d.prtimesheetid IS NULL THEN 1 + ELSE d.prtimesheetid + END,',', + CASE + WHEN d.prassignmentid IS NULL THEN 1 + ELSE d.prassignmentid + END,',', + CASE + WHEN d.prchargecodeid IS NULL THEN 1 + ELSE d.prchargecodeid + END,',', + CASE + WHEN (d.prtypecodeid) IS NULL THEN '' + ELSE d.prtypecodeid + END,',', + CASE + WHEN d.practsum IS NULL THEN 1 + ELSE d.practsum + END,',', + CASE + WHEN d.prsequence IS NULL THEN 1 + ELSE d.prsequence + END,',', + CASE + WHEN length(d.prmodby) IS NULL THEN '' + ELSE d.prmodby + END,',', + CASE + WHEN d.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) + ELSE d.prmodtime + END,',', + CASE + WHEN d.prrmexported IS NULL THEN 1 + ELSE d.prrmexported + END,',', + CASE + WHEN d.prrmckdel IS NULL THEN 1 + ELSE d.prrmckdel + END,',', + CASE + WHEN d.slice_status IS NULL THEN 1 + ELSE d.slice_status + END,',', + CASE + WHEN d.role_id IS NULL THEN 1 + ELSE d.role_id + END,',', + CASE + WHEN length(d.user_lov1) IS NULL THEN '' + ELSE d.user_lov1 + END,',', + CASE + WHEN length(d.user_lov2) IS NULL THEN '' + ELSE d.user_lov2 + END,',', + CASE + WHEN d.incident_id IS NULL THEN 1 + ELSE d.incident_id + END,',', + CASE + WHEN d.incident_investment_id IS NULL THEN 1 + ELSE d.incident_investment_id + END,',', + CASE + WHEN d.odf_ss_actuals IS NULL THEN 1 + ELSE d.odf_ss_actuals + END ) +PREHOOK: type: QUERY +PREHOOK: Input: default@table1_n8 +PREHOOK: Input: default@table2_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain SELECT s.idp_warehouse_id AS source_warehouse_id + FROM table1_n8 s + JOIN + + table2_n4 d + ON ( + s.prid = d.prid ) + JOIN + table2_n4 e + ON + s.prid = e.prid + WHERE + concat( + CASE + WHEN s.prid IS NULL THEN 1 + ELSE s.prid + END,',', + CASE + WHEN s.prtimesheetid IS NULL THEN 1 + ELSE s.prtimesheetid + END,',', + CASE + WHEN s.prassignmentid IS NULL THEN 1 + ELSE s.prassignmentid + END,',', + CASE + WHEN s.prchargecodeid IS NULL THEN 1 + ELSE s.prchargecodeid + END,',', + CASE + WHEN (s.prtypecodeid) IS NULL THEN '' + ELSE s.prtypecodeid + END,',', + CASE + WHEN s.practsum IS NULL THEN 1 + ELSE s.practsum + END,',', + CASE + WHEN s.prsequence IS NULL THEN 1 + ELSE s.prsequence + END,',', + CASE + WHEN length(s.prmodby) IS NULL THEN '' + ELSE s.prmodby + END,',', + CASE + WHEN s.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) + ELSE s.prmodtime + END,',', + CASE + WHEN s.prrmexported IS NULL THEN 1 + ELSE s.prrmexported + END,',', + CASE + WHEN s.prrmckdel IS NULL THEN 1 + ELSE s.prrmckdel + END,',', + CASE + WHEN s.slice_status IS NULL THEN 1 + ELSE s.slice_status + END,',', + CASE + WHEN s.role_id IS NULL THEN 1 + ELSE s.role_id + END,',', + CASE + WHEN length(s.user_lov1) IS NULL THEN '' + ELSE s.user_lov1 + END,',', + CASE + WHEN length(s.user_lov2) IS NULL THEN '' + ELSE s.user_lov2 + END,',', + CASE + WHEN s.incident_id IS NULL THEN 1 + ELSE s.incident_id + END,',', + CASE + WHEN s.incident_investment_id IS NULL THEN 1 + ELSE s.incident_investment_id + END,',', + CASE + WHEN s.odf_ss_actuals IS NULL THEN 1 + ELSE s.odf_ss_actuals + END ) != concat( + CASE + WHEN length(d.pruid) IS NULL THEN '' + ELSE d.pruid + END,',', + CASE + WHEN d.prid IS NULL THEN 1 + ELSE d.prid + END,',', + CASE + WHEN d.prtimesheetid IS NULL THEN 1 + ELSE d.prtimesheetid + END,',', + CASE + WHEN d.prassignmentid IS NULL THEN 1 + ELSE d.prassignmentid + END,',', + CASE + WHEN d.prchargecodeid IS NULL THEN 1 + ELSE d.prchargecodeid + END,',', + CASE + WHEN (d.prtypecodeid) IS NULL THEN '' + ELSE d.prtypecodeid + END,',', + CASE + WHEN d.practsum IS NULL THEN 1 + ELSE d.practsum + END,',', + CASE + WHEN d.prsequence IS NULL THEN 1 + ELSE d.prsequence + END,',', + CASE + WHEN length(d.prmodby) IS NULL THEN '' + ELSE d.prmodby + END,',', + CASE + WHEN d.prmodtime IS NULL THEN cast(from_unixtime(unix_timestamp('2017-12-08','yyyy-MM-dd') ) AS timestamp) + ELSE d.prmodtime + END,',', + CASE + WHEN d.prrmexported IS NULL THEN 1 + ELSE d.prrmexported + END,',', + CASE + WHEN d.prrmckdel IS NULL THEN 1 + ELSE d.prrmckdel + END,',', + CASE + WHEN d.slice_status IS NULL THEN 1 + ELSE d.slice_status + END,',', + CASE + WHEN d.role_id IS NULL THEN 1 + ELSE d.role_id + END,',', + CASE + WHEN length(d.user_lov1) IS NULL THEN '' + ELSE d.user_lov1 + END,',', + CASE + WHEN length(d.user_lov2) IS NULL THEN '' + ELSE d.user_lov2 + END,',', + CASE + WHEN d.incident_id IS NULL THEN 1 + ELSE d.incident_id + END,',', + CASE + WHEN d.incident_investment_id IS NULL THEN 1 + ELSE d.incident_investment_id + END,',', + CASE + WHEN d.odf_ss_actuals IS NULL THEN 1 + ELSE d.odf_ss_actuals + END ) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@table1_n8 +POSTHOOK: Input: default@table2_n4 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s + filterExpr: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: idp_warehouse_id (type: bigint), prid (type: bigint), concat(CAST( CASE WHEN (prid is null) THEN (1L) ELSE (prid) END AS STRING), ',', CASE WHEN (prtimesheetid is null) THEN (1L) ELSE (prtimesheetid) END, ',', CASE WHEN (prassignmentid is null) THEN (1L) ELSE (prassignmentid) END, ',', CASE WHEN (prchargecodeid is null) THEN (1L) ELSE (prchargecodeid) END, ',', CASE WHEN (prtypecodeid is null) THEN ('') ELSE (CAST( prtypecodeid AS STRING)) END, ',', CASE WHEN (practsum is null) THEN (1) ELSE (practsum) END, ',', CASE WHEN (prsequence is null) THEN (1L) ELSE (prsequence) END, ',', CASE WHEN (length(prmodby) is null) THEN ('') ELSE (prmodby) END, ',', CASE WHEN (prmodtime is null) THEN (TIMESTAMP'2017-12-08 00:00:00') ELSE (prmodtime) END, ',', CASE WHEN (prrmexported is null) THEN (1L) ELSE (prrmexported) END, ',', CASE WHEN (prrmckdel is null) THEN (1L) ELSE (prrmckdel) END, ',', CASE WHEN (slice_status is null) THEN (1) ELSE (slice_status) END, ',', CASE WHEN (role_id is null) THEN (1L) ELSE (role_id) END, ',', CASE WHEN (length(user_lov1) is null) THEN ('') ELSE (user_lov1) END, ',', CASE WHEN (length(user_lov2) is null) THEN ('') ELSE (user_lov2) END, ',', CASE WHEN (incident_id is null) THEN (1L) ELSE (incident_id) END, ',', CASE WHEN (incident_investment_id is null) THEN (1L) ELSE (incident_investment_id) END, ',', CASE WHEN (odf_ss_actuals is null) THEN (1L) ELSE (odf_ss_actuals) END) (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: bigint) + Statistics: Num rows: 1 Data size: 668 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: bigint), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: e + filterExpr: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: prid (type: bigint) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: bigint) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: d + filterExpr: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: prid is not null (type: boolean) + Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: prid (type: bigint), concat(CASE WHEN (length(pruid) is null) THEN ('') ELSE (pruid) END, ',', CASE WHEN (prid is null) THEN (1L) ELSE (prid) END, ',', CASE WHEN (prtimesheetid is null) THEN (1L) ELSE (prtimesheetid) END, ',', CASE WHEN (prassignmentid is null) THEN (1L) ELSE (prassignmentid) END, ',', CASE WHEN (prchargecodeid is null) THEN (1L) ELSE (prchargecodeid) END, ',', CASE WHEN (prtypecodeid is null) THEN ('') ELSE (CAST( prtypecodeid AS STRING)) END, ',', CASE WHEN (practsum is null) THEN (1) ELSE (practsum) END, ',', CASE WHEN (prsequence is null) THEN (1L) ELSE (prsequence) END, ',', CASE WHEN (length(prmodby) is null) THEN ('') ELSE (prmodby) END, ',', CASE WHEN (prmodtime is null) THEN (TIMESTAMP'2017-12-08 00:00:00') ELSE (prmodtime) END, ',', CASE WHEN (prrmexported is null) THEN (1L) ELSE (prrmexported) END, ',', CASE WHEN (prrmckdel is null) THEN (1L) ELSE (prrmckdel) END, ',', CASE WHEN (slice_status is null) THEN (1) ELSE (slice_status) END, ',', CASE WHEN (role_id is null) THEN (1L) ELSE (role_id) END, ',', CASE WHEN (length(user_lov1) is null) THEN ('') ELSE (user_lov1) END, ',', CASE WHEN (length(user_lov2) is null) THEN ('') ELSE (user_lov2) END, ',', CASE WHEN (incident_id is null) THEN (1L) ELSE (incident_id) END, ',', CASE WHEN (incident_investment_id is null) THEN (1L) ELSE (incident_investment_id) END, ',', CASE WHEN (odf_ss_actuals is null) THEN (1L) ELSE (odf_ss_actuals) END) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: bigint) + Statistics: Num rows: 1 Data size: 776 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: bigint) + 1 _col0 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 734 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: bigint) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: bigint) + Statistics: Num rows: 1 Data size: 734 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: bigint), _col2 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: bigint) + 1 _col0 (type: bigint) + outputColumnNames: _col0, _col2, _col5 + residual filter predicates: {(_col2 <> _col5)} + Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: bigint) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 807 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: drop table table2_n4 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@table2_n4 +PREHOOK: Output: default@table2_n4 +POSTHOOK: query: drop table table2_n4 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@table2_n4 +POSTHOOK: Output: default@table2_n4 +PREHOOK: query: drop table table1_n8 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@table1_n8 +PREHOOK: Output: default@table1_n8 +POSTHOOK: query: drop table table1_n8 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@table1_n8 +POSTHOOK: Output: default@table1_n8 diff --git ql/src/test/results/clientpositive/innerjoin.q.out ql/src/test/results/clientpositive/llap/innerjoin.q.out similarity index 66% rename from ql/src/test/results/clientpositive/innerjoin.q.out rename to ql/src/test/results/clientpositive/llap/innerjoin.q.out index 10c15e6fac..5355097c9c 100644 --- ql/src/test/results/clientpositive/innerjoin.q.out +++ ql/src/test/results/clientpositive/llap/innerjoin.q.out @@ -20,86 +20,118 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@dest_j1_n20 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n20 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n20 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -111,7 +143,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n20 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -119,30 +151,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest_j1_n20 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM src src1 INNER JOIN src src2 ON (src1.key = src2.key) INSERT OVERWRITE TABLE dest_j1_n20 SELECT src1.key, src2.value PREHOOK: type: QUERY @@ -1239,56 +1247,70 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/innerjoin1.q.out ql/src/test/results/clientpositive/llap/innerjoin1.q.out new file mode 100644 index 0000000000..a5d94a971d --- /dev/null +++ ql/src/test/results/clientpositive/llap/innerjoin1.q.out @@ -0,0 +1,278 @@ +PREHOOK: query: DROP TABLE IF EXISTS innerjoinsrc1 +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE IF EXISTS innerjoinsrc1 +POSTHOOK: type: DROPTABLE +PREHOOK: query: DROP TABLE IF EXISTS innerjoinsrc2 +PREHOOK: type: DROPTABLE +POSTHOOK: query: DROP TABLE IF EXISTS innerjoinsrc2 +POSTHOOK: type: DROPTABLE +PREHOOK: query: CREATE TABLE innerjoinsrc1 (a1 STRING, a2 STRING) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@innerjoinsrc1 +POSTHOOK: query: CREATE TABLE innerjoinsrc1 (a1 STRING, a2 STRING) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@innerjoinsrc1 +PREHOOK: query: CREATE TABLE innerjoinsrc2 (b1 STRING, b2 STRING) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@innerjoinsrc2 +POSTHOOK: query: CREATE TABLE innerjoinsrc2 (b1 STRING, b2 STRING) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@innerjoinsrc2 +PREHOOK: query: INSERT INTO TABLE innerjoinsrc1 (a1) VALUES ('1') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@innerjoinsrc1 +POSTHOOK: query: INSERT INTO TABLE innerjoinsrc1 (a1) VALUES ('1') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@innerjoinsrc1 +POSTHOOK: Lineage: innerjoinsrc1.a1 SCRIPT [] +POSTHOOK: Lineage: innerjoinsrc1.a2 SIMPLE [] +PREHOOK: query: INSERT INTO TABLE innerjoinsrc1 VALUES ('2', '2') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@innerjoinsrc1 +POSTHOOK: query: INSERT INTO TABLE innerjoinsrc1 VALUES ('2', '2') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@innerjoinsrc1 +POSTHOOK: Lineage: innerjoinsrc1.a1 SCRIPT [] +POSTHOOK: Lineage: innerjoinsrc1.a2 SCRIPT [] +PREHOOK: query: INSERT INTO TABLE innerjoinsrc2 (b1) VALUES ('1') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@innerjoinsrc2 +POSTHOOK: query: INSERT INTO TABLE innerjoinsrc2 (b1) VALUES ('1') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@innerjoinsrc2 +POSTHOOK: Lineage: innerjoinsrc2.b1 SCRIPT [] +POSTHOOK: Lineage: innerjoinsrc2.b2 SIMPLE [] +PREHOOK: query: INSERT INTO TABLE innerjoinsrc2 VALUES ('2', '2') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@innerjoinsrc2 +POSTHOOK: query: INSERT INTO TABLE innerjoinsrc2 VALUES ('2', '2') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@innerjoinsrc2 +POSTHOOK: Lineage: innerjoinsrc2.b1 SCRIPT [] +POSTHOOK: Lineage: innerjoinsrc2.b2 SCRIPT [] +PREHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) +PREHOOK: type: QUERY +PREHOOK: Input: default@innerjoinsrc1 +PREHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@innerjoinsrc1 +POSTHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: c1 + filterExpr: (a1 is not null or a2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (a1 is not null or a2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: a1 (type: string), a2 (type: string), CASE WHEN (a1 is not null) THEN (a1) ELSE (a2) END (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: c2 + filterExpr: (b1 is not null or b2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (b1 is not null or b2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: b1 (type: string), b2 (type: string), CASE WHEN (b1 is not null) THEN (b1) ELSE (b2) END (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: string) + 1 _col2 (type: string) + outputColumnNames: _col0, _col1, _col3, _col4 + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) +PREHOOK: type: QUERY +PREHOOK: Input: default@innerjoinsrc1 +PREHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM innerjoinsrc1 c1, innerjoinsrc2 c2 WHERE COALESCE(a1,a2)=COALESCE(b1,b2) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@innerjoinsrc1 +POSTHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +1 NULL 1 NULL +2 2 2 2 +PREHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) +PREHOOK: type: QUERY +PREHOOK: Input: default@innerjoinsrc1 +PREHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@innerjoinsrc1 +POSTHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: c1 + filterExpr: (a1 is not null or a2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (a1 is not null or a2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: a1 (type: string), a2 (type: string), CASE WHEN (a1 is not null) THEN (a1) ELSE (a2) END (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: c2 + filterExpr: (b1 is not null or b2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (b1 is not null or b2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: b1 (type: string), b2 (type: string), CASE WHEN (b1 is not null) THEN (b1) ELSE (b2) END (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 708 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: string) + 1 _col2 (type: string) + outputColumnNames: _col0, _col1, _col3, _col4 + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 680 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) +PREHOOK: type: QUERY +PREHOOK: Input: default@innerjoinsrc1 +PREHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM innerjoinsrc1 c1 inner join innerjoinsrc2 c2 ON (COALESCE(a1,a2)=COALESCE(b1,b2)) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@innerjoinsrc1 +POSTHOOK: Input: default@innerjoinsrc2 +#### A masked pattern was here #### +1 NULL 1 NULL +2 2 2 2 diff --git ql/src/test/results/clientpositive/input.q.out ql/src/test/results/clientpositive/llap/input.q.out similarity index 97% rename from ql/src/test/results/clientpositive/input.q.out rename to ql/src/test/results/clientpositive/llap/input.q.out index 46efea5663..d820ed61e1 100644 --- ql/src/test/results/clientpositive/input.q.out +++ ql/src/test/results/clientpositive/llap/input.q.out @@ -18,11 +18,9 @@ STAGE PLANS: Processor Tree: TableScan alias: x - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRC x diff --git ql/src/test/results/clientpositive/input0.q.out ql/src/test/results/clientpositive/llap/input0.q.out similarity index 97% rename from ql/src/test/results/clientpositive/input0.q.out rename to ql/src/test/results/clientpositive/llap/input0.q.out index d6d3f7415b..1ba9e95cc7 100644 --- ql/src/test/results/clientpositive/input0.q.out +++ ql/src/test/results/clientpositive/llap/input0.q.out @@ -18,11 +18,9 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT * FROM src diff --git ql/src/test/results/clientpositive/llap/input11.q.out ql/src/test/results/clientpositive/llap/input11.q.out new file mode 100644 index 0000000000..6b6415f375 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input11.q.out @@ -0,0 +1,213 @@ +PREHOOK: query: CREATE TABLE dest1_n152(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n152 +POSTHOOK: query: CREATE TABLE dest1_n152(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n152 +PREHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n152 +POSTHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n152 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n152 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n152 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.dest1_n152 + +PREHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n152 +POSTHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n152 SELECT src.key, src.value WHERE src.key < 100 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n152 +POSTHOOK: Lineage: dest1_n152.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n152.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n152.* FROM dest1_n152 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n152 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n152.* FROM dest1_n152 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n152 +#### A masked pattern was here #### +86 val_86 +27 val_27 +98 val_98 +66 val_66 +37 val_37 +15 val_15 +82 val_82 +17 val_17 +0 val_0 +57 val_57 +20 val_20 +92 val_92 +47 val_47 +72 val_72 +4 val_4 +35 val_35 +54 val_54 +51 val_51 +65 val_65 +83 val_83 +12 val_12 +67 val_67 +84 val_84 +58 val_58 +8 val_8 +24 val_24 +42 val_42 +0 val_0 +96 val_96 +26 val_26 +51 val_51 +43 val_43 +95 val_95 +98 val_98 +85 val_85 +77 val_77 +0 val_0 +87 val_87 +15 val_15 +72 val_72 +90 val_90 +19 val_19 +10 val_10 +5 val_5 +58 val_58 +35 val_35 +95 val_95 +11 val_11 +34 val_34 +42 val_42 +78 val_78 +76 val_76 +41 val_41 +30 val_30 +64 val_64 +76 val_76 +74 val_74 +69 val_69 +33 val_33 +70 val_70 +5 val_5 +2 val_2 +35 val_35 +80 val_80 +44 val_44 +53 val_53 +90 val_90 +12 val_12 +5 val_5 +70 val_70 +24 val_24 +70 val_70 +83 val_83 +26 val_26 +67 val_67 +18 val_18 +9 val_9 +18 val_18 +97 val_97 +84 val_84 +28 val_28 +37 val_37 +90 val_90 +97 val_97 diff --git ql/src/test/results/clientpositive/llap/input11_limit.q.out ql/src/test/results/clientpositive/llap/input11_limit.q.out new file mode 100644 index 0000000000..5971243345 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input11_limit.q.out @@ -0,0 +1,128 @@ +PREHOOK: query: CREATE TABLE dest1_n153(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n153 +POSTHOOK: query: CREATE TABLE dest1_n153(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n153 +PREHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n153 +POSTHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n153 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n153 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n153 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + +PREHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n153 +POSTHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n153 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n153 +POSTHOOK: Lineage: dest1_n153.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n153.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n153.* FROM dest1_n153 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n153 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n153.* FROM dest1_n153 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n153 +#### A masked pattern was here #### +0 val_0 +15 val_15 +17 val_17 +27 val_27 +37 val_37 +57 val_57 +66 val_66 +82 val_82 +86 val_86 +98 val_98 diff --git ql/src/test/results/clientpositive/input12.q.out ql/src/test/results/clientpositive/llap/input12.q.out similarity index 58% rename from ql/src/test/results/clientpositive/input12.q.out rename to ql/src/test/results/clientpositive/llap/input12.q.out index 0da1c6e33a..66ac09aa05 100644 --- ql/src/test/results/clientpositive/input12.q.out +++ ql/src/test/results/clientpositive/llap/input12.q.out @@ -44,142 +44,177 @@ POSTHOOK: Output: default@dest2_n32 POSTHOOK: Output: default@dest3_n5@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-3 is a root stage - Stage-9 depends on stages: Stage-3 , consists of Stage-6, Stage-5, Stage-7 - Stage-6 - Stage-0 depends on stages: Stage-6, Stage-5, Stage-8 - Stage-4 depends on stages: Stage-0, Stage-11, Stage-13 - Stage-5 - Stage-7 - Stage-8 depends on stages: Stage-7 - Stage-1 depends on stages: Stage-3 - Stage-10 depends on stages: Stage-1, Stage-11, Stage-13 - Stage-11 depends on stages: Stage-3 - Stage-12 depends on stages: Stage-2, Stage-11, Stage-13 - Stage-2 depends on stages: Stage-3 - Stage-13 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-4 + Stage-6 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-4 + Stage-7 depends on stages: Stage-2 STAGE PLANS: Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 100) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 100) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n122 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: ((key >= 100) and (key < 200)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n32 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9818182 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: (key >= 200) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 664 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 664 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest3_n5 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: key + Statistics: Num rows: 166 Data size: 30544 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + keys: '2008-04-08' (type: string), '12' (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: '2008-04-08' (type: string), '12' (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: '2008-04-08' (type: string), '12' (type: string) + Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n122 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Filter Operator - predicate: ((key >= 100) and (key < 200)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n32 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (key >= 200) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 664 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 166 Data size: 664 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest3_n5 + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + keys: '2008-04-08' (type: string), '12' (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: _col0 (type: int) - outputColumnNames: key - Statistics: Num rows: 166 Data size: 30544 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll') - keys: '2008-04-08' (type: string), '12' (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + expressions: _col2 (type: struct), '2008-04-08' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-9 - Conditional Operator - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-4 + Dependency Collection Stage: Stage-0 Move Operator @@ -191,7 +226,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n122 - Stage: Stage-4 + Stage: Stage-5 Stats Work Basic Stats Work: Column Stats Desc: @@ -199,36 +234,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n122 - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n122 - - Stage: Stage-7 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n122 - - Stage: Stage-8 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - Stage: Stage-1 Move Operator tables: @@ -239,7 +244,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n32 - Stage: Stage-10 + Stage: Stage-6 Stats Work Basic Stats Work: Column Stats Desc: @@ -247,38 +252,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest2_n32 - Stage: Stage-11 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-12 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key - Column Types: int - Table: default.dest3_n5 - Stage: Stage-2 Move Operator tables: @@ -292,36 +265,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest3_n5 - Stage: Stage-13 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: '2008-04-08' (type: string), '12' (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: '2008-04-08' (type: string), '12' (type: string) - Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - keys: '2008-04-08' (type: string), '12' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), '2008-04-08' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-7 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key + Column Types: int + Table: default.dest3_n5 PREHOOK: query: FROM src INSERT OVERWRITE TABLE dest1_n122 SELECT src.* WHERE src.key < 100 diff --git ql/src/test/results/clientpositive/input13.q.out ql/src/test/results/clientpositive/llap/input13.q.out similarity index 54% rename from ql/src/test/results/clientpositive/input13.q.out rename to ql/src/test/results/clientpositive/llap/input13.q.out index f38eeace1c..62b0e61daf 100644 --- ql/src/test/results/clientpositive/input13.q.out +++ ql/src/test/results/clientpositive/llap/input13.q.out @@ -48,157 +48,192 @@ POSTHOOK: Output: default@dest3_n1@ds=2008-04-08/hr=12 POSTHOOK: Output: target/warehouse/dest4.out STAGE DEPENDENCIES: Stage-4 is a root stage - Stage-10 depends on stages: Stage-4 , consists of Stage-7, Stage-6, Stage-8 - Stage-7 - Stage-0 depends on stages: Stage-7, Stage-6, Stage-9 - Stage-5 depends on stages: Stage-0, Stage-12, Stage-14, Stage-3 - Stage-6 - Stage-8 - Stage-9 depends on stages: Stage-8 - Stage-1 depends on stages: Stage-4 - Stage-11 depends on stages: Stage-1, Stage-12, Stage-14, Stage-3 - Stage-12 depends on stages: Stage-4 - Stage-13 depends on stages: Stage-2, Stage-12, Stage-14, Stage-3 - Stage-2 depends on stages: Stage-4 - Stage-14 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-5 + Stage-6 depends on stages: Stage-0, Stage-3 + Stage-1 depends on stages: Stage-5 + Stage-7 depends on stages: Stage-1, Stage-3 + Stage-2 depends on stages: Stage-5 + Stage-8 depends on stages: Stage-2, Stage-3 Stage-3 depends on stages: Stage-4 STAGE PLANS: Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key < 100) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 100) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n77 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: ((key >= 100) and (key < 200)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n16 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9818182 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Filter Operator + predicate: ((key >= 200) and (key < 300)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 55 Data size: 220 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 220 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest3_n1 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: key + Statistics: Num rows: 55 Data size: 10120 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + keys: '2008-04-08' (type: string), '12' (type: string) + minReductionHashAggr: 0.9818182 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: '2008-04-08' (type: string), '12' (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: '2008-04-08' (type: string), '12' (type: string) + Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct) + Filter Operator + predicate: (key >= 300) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 15106 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 15106 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n77 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 166 Data size: 15770 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Filter Operator - predicate: ((key >= 100) and (key < 200)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest2_n16 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 55 Data size: 5225 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: ((key >= 200) and (key < 300)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 55 Data size: 220 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 55 Data size: 220 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest3_n1 + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + keys: '2008-04-08' (type: string), '12' (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: _col0 (type: int) - outputColumnNames: key - Statistics: Num rows: 55 Data size: 10120 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll') - keys: '2008-04-08' (type: string), '12' (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Filter Operator - predicate: (key >= 300) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 15106 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 166 Data size: 15106 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + expressions: _col2 (type: struct), '2008-04-08' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-10 - Conditional Operator - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-5 + Dependency Collection Stage: Stage-0 Move Operator @@ -210,7 +245,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n77 - Stage: Stage-5 + Stage: Stage-6 Stats Work Basic Stats Work: Column Stats Desc: @@ -218,36 +253,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n77 - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n77 - - Stage: Stage-8 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n77 - - Stage: Stage-9 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - Stage: Stage-1 Move Operator tables: @@ -258,7 +263,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n16 - Stage: Stage-11 + Stage: Stage-7 Stats Work Basic Stats Work: Column Stats Desc: @@ -266,38 +271,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest2_n16 - Stage: Stage-12 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-13 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key - Column Types: int - Table: default.dest3_n1 - Stage: Stage-2 Move Operator tables: @@ -311,36 +284,13 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest3_n1 - Stage: Stage-14 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: '2008-04-08' (type: string), '12' (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: '2008-04-08' (type: string), '12' (type: string) - Statistics: Num rows: 1 Data size: 604 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - keys: '2008-04-08' (type: string), '12' (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), '2008-04-08' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 620 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Stage: Stage-8 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key + Column Types: int + Table: default.dest3_n1 Stage: Stage-3 Move Operator diff --git ql/src/test/results/clientpositive/llap/input1_limit.q.out ql/src/test/results/clientpositive/llap/input1_limit.q.out new file mode 100644 index 0000000000..0e9d0dec21 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input1_limit.q.out @@ -0,0 +1,283 @@ +PREHOOK: query: CREATE TABLE dest1_n12(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n12 +POSTHOOK: query: CREATE TABLE dest1_n12(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n12 +PREHOOK: query: CREATE TABLE dest2_n2(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest2_n2 +POSTHOOK: query: CREATE TABLE dest2_n2(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest2_n2 +PREHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n12 +PREHOOK: Output: default@dest2_n2 +POSTHOOK: query: EXPLAIN +FROM src +INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n12 +POSTHOOK: Output: default@dest2_n2 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key < 100) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string) + Filter Operator + predicate: (key < 100) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n12 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 10 Data size: 950 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n2 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.8 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n12 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.dest1_n12 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n2 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.dest2_n2 + +PREHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n12 +PREHOOK: Output: default@dest2_n2 +POSTHOOK: query: FROM src +INSERT OVERWRITE TABLE dest1_n12 SELECT src.key, src.value WHERE src.key < 100 LIMIT 10 +INSERT OVERWRITE TABLE dest2_n2 SELECT src.key, src.value WHERE src.key < 100 LIMIT 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n12 +POSTHOOK: Output: default@dest2_n2 +POSTHOOK: Lineage: dest1_n12.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n12.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n12.* FROM dest1_n12 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n12 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n12.* FROM dest1_n12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n12 +#### A masked pattern was here #### +0 val_0 +15 val_15 +17 val_17 +27 val_27 +37 val_37 +57 val_57 +66 val_66 +82 val_82 +86 val_86 +98 val_98 +PREHOOK: query: SELECT dest2_n2.* FROM dest2_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest2_n2.* FROM dest2_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n2 +#### A masked pattern was here #### +27 val_27 +37 val_37 +66 val_66 +86 val_86 +98 val_98 diff --git ql/src/test/results/clientpositive/llap/input21.q.out ql/src/test/results/clientpositive/llap/input21.q.out new file mode 100644 index 0000000000..d87336a10f --- /dev/null +++ ql/src/test/results/clientpositive/llap/input21.q.out @@ -0,0 +1,93 @@ +PREHOOK: query: CREATE TABLE src_null(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@src_null +POSTHOOK: query: CREATE TABLE src_null(a STRING, b STRING, c STRING, d STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@src_null +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE src_null +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@src_null +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/null.txt' INTO TABLE src_null +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@src_null +PREHOOK: query: EXPLAIN SELECT * FROM src_null DISTRIBUTE BY c SORT BY d +PREHOOK: type: QUERY +PREHOOK: Input: default@src_null +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src_null DISTRIBUTE BY c SORT BY d +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_null +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_null + Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string), c (type: string), d (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col3 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string), KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 736 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM src_null DISTRIBUTE BY c SORT BY d +PREHOOK: type: QUERY +PREHOOK: Input: default@src_null +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM src_null DISTRIBUTE BY c SORT BY d +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_null +#### A masked pattern was here #### +1.0 1 same 0 +1.0 1 same 1 +1.0 1 same 2 +1.0 1 same 3 +1.0 1 same 4 +NULL 1 same 5 +NULL NULL same 6 +1.0 NULL same 7 +1.0 1 same 8 +1.0 1 same 9 diff --git ql/src/test/results/clientpositive/llap/input22.q.out ql/src/test/results/clientpositive/llap/input22.q.out new file mode 100644 index 0000000000..832aaeebb5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input22.q.out @@ -0,0 +1,115 @@ +PREHOOK: query: CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@INPUT4 +POSTHOOK: query: CREATE TABLE INPUT4(KEY STRING, VALUE STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@INPUT4 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE INPUT4 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@input4 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE INPUT4 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@input4 +PREHOOK: query: EXPLAIN +SELECT a.KEY2 +FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 + FROM INPUT4) a +ORDER BY KEY2 LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@input4 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT a.KEY2 +FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 + FROM INPUT4) a +ORDER BY KEY2 LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@input4 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: input4 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: + + keys: key (type: string) + null sort order: z + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + top n: 10 + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT a.KEY2 +FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 + FROM INPUT4) a +ORDER BY KEY2 LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@input4 +#### A masked pattern was here #### +POSTHOOK: query: SELECT a.KEY2 +FROM (SELECT INPUT4.*, INPUT4.KEY as KEY2 + FROM INPUT4) a +ORDER BY KEY2 LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@input4 +#### A masked pattern was here #### +0 +0 +0 +10 +100 +100 +103 +103 +104 +104 diff --git ql/src/test/results/clientpositive/llap/input23.q.out ql/src/test/results/clientpositive/llap/input23.q.out new file mode 100644 index 0000000000..e0799d4d07 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input23.q.out @@ -0,0 +1,197 @@ +Warning: Shuffle Join MERGEJOIN[14][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: explain extended + select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: explain extended + select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, CAST('2008-04-08' AS STRING) AS `ds`, CAST('11' AS STRING) AS `hr`, `t2`.`key` AS `key1`, `t2`.`value` AS `value1`, CAST('2008-04-08' AS STRING) AS `ds1`, CAST('14' AS STRING) AS `hr1` +FROM (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' AND `hr` = '11') AS `t0`, +(SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' AND `hr` = '14') AS `t2` +LIMIT 5 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((ds = '2008-04-08') and (hr = '11')) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [a] + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) + Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) + Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false + Execution mode: vectorized, llap + LLAP IO: unknown + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Statistics: Num rows: 500 Data size: 173000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string), _col2 (type: string), _col3 (type: string), '2008-04-08' (type: string), '14' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 500 Data size: 353000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 5 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 + columns.types string:string:string:string:string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: 5 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[14][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: select * from srcpart a join srcpart b where a.ds = '2008-04-08' and a.hr = '11' and b.ds = '2008-04-08' and b.hr = '14' limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/llap/input25.q.out ql/src/test/results/clientpositive/llap/input25.q.out new file mode 100644 index 0000000000..0dd79dd18e --- /dev/null +++ ql/src/test/results/clientpositive/llap/input25.q.out @@ -0,0 +1,170 @@ +PREHOOK: query: create table tst(a int, b int) partitioned by (d string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tst +POSTHOOK: query: create table tst(a int, b int) partitioned by (d string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tst +PREHOOK: query: alter table tst add partition (d='2009-01-01') +PREHOOK: type: ALTERTABLE_ADDPARTS +PREHOOK: Output: default@tst +POSTHOOK: query: alter table tst add partition (d='2009-01-01') +POSTHOOK: type: ALTERTABLE_ADDPARTS +POSTHOOK: Output: default@tst +POSTHOOK: Output: default@tst@d=2009-01-01 +PREHOOK: query: alter table tst add partition (d='2009-02-02') +PREHOOK: type: ALTERTABLE_ADDPARTS +PREHOOK: Output: default@tst +POSTHOOK: query: alter table tst add partition (d='2009-02-02') +POSTHOOK: type: ALTERTABLE_ADDPARTS +POSTHOOK: Output: default@tst +POSTHOOK: Output: default@tst@d=2009-02-02 +PREHOOK: query: explain +select * from ( + select * from (select * from tst x where x.d='2009-01-01' limit 10)a + union all + select * from (select * from tst x where x.d='2009-02-02' limit 10)b +) subq +PREHOOK: type: QUERY +PREHOOK: Input: default@tst +PREHOOK: Input: default@tst@d=2009-01-01 +PREHOOK: Input: default@tst@d=2009-02-02 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from ( + select * from (select * from tst x where x.d='2009-01-01' limit 10)a + union all + select * from (select * from tst x where x.d='2009-02-02' limit 10)b +) subq +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tst +POSTHOOK: Input: default@tst@d=2009-01-01 +POSTHOOK: Input: default@tst@d=2009-02-02 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) + Reducer 5 <- Map 4 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: (d = '2009-01-01') (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: a (type: int), b (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: int), _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: x + filterExpr: (d = '2009-02-02') (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: a (type: int), b (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: int), _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), '2009-01-01' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 10 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), '2009-02-02' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 16 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 3 + Vertex: Union 3 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from ( + select * from (select * from tst x where x.d='2009-01-01' limit 10)a + union all + select * from (select * from tst x where x.d='2009-02-02' limit 10)b +) subq +PREHOOK: type: QUERY +PREHOOK: Input: default@tst +PREHOOK: Input: default@tst@d=2009-01-01 +PREHOOK: Input: default@tst@d=2009-02-02 +#### A masked pattern was here #### +POSTHOOK: query: select * from ( + select * from (select * from tst x where x.d='2009-01-01' limit 10)a + union all + select * from (select * from tst x where x.d='2009-02-02' limit 10)b +) subq +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tst +POSTHOOK: Input: default@tst@d=2009-01-01 +POSTHOOK: Input: default@tst@d=2009-02-02 +#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/llap/input26.q.out ql/src/test/results/clientpositive/llap/input26.q.out new file mode 100644 index 0000000000..6d563a314a --- /dev/null +++ ql/src/test/results/clientpositive/llap/input26.q.out @@ -0,0 +1,164 @@ +PREHOOK: query: explain +select * from ( + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa + union all + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb +)subq +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from ( + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa + union all + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb +)subq +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Union 3 (CONTAINS) + Reducer 5 <- Map 4 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((ds = '2008-04-08') and (hr = '11')) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: key (type: string) + null sort order: z + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 5 + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) + Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((ds = '2008-04-08') and (hr = '14')) (type: boolean) + Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: unknown + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 890 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 5 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 5 + Statistics: Num rows: 1 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '14' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 254 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 2148 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 3 + Vertex: Union 3 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from ( + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa + union all + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb +)subq +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: select * from ( + select * from (select * from srcpart a where a.ds = '2008-04-08' and a.hr = '11' order by a.key limit 5)pa + union all + select * from (select * from srcpart b where b.ds = '2008-04-08' and b.hr = '14' limit 5)pb +)subq +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +10 val_10 2008-04-08 11 +100 val_100 2008-04-08 11 diff --git ql/src/test/results/clientpositive/llap/input2_limit.q.out ql/src/test/results/clientpositive/llap/input2_limit.q.out new file mode 100644 index 0000000000..f4b5b0ab9b --- /dev/null +++ ql/src/test/results/clientpositive/llap/input2_limit.q.out @@ -0,0 +1,43 @@ +PREHOOK: query: EXPLAIN +SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 5 + Processor Tree: + TableScan + alias: x + filterExpr: (UDFToDouble(key) < 300.0D) (type: boolean) + Filter Operator + predicate: (UDFToDouble(key) < 300.0D) (type: boolean) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 5 + ListSink + +PREHOOK: query: SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT x.* FROM SRC x WHERE x.key < 300 LIMIT 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +238 val_238 +86 val_86 +27 val_27 +165 val_165 +255 val_255 diff --git ql/src/test/results/clientpositive/llap/input30.q.out ql/src/test/results/clientpositive/llap/input30.q.out new file mode 100644 index 0000000000..769df05dac --- /dev/null +++ ql/src/test/results/clientpositive/llap/input30.q.out @@ -0,0 +1,269 @@ +PREHOOK: query: create table dest30(a int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest30 +POSTHOOK: query: create table dest30(a int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest30 +PREHOOK: query: create table tst_dest30(a int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tst_dest30 +POSTHOOK: query: create table tst_dest30(a int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tst_dest30 +PREHOOK: query: explain +insert overwrite table dest30 +select count(1) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tst_dest30 +POSTHOOK: query: explain +insert overwrite table dest30 +select count(1) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tst_dest30 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((hash(rand(460476415)) & 2147483647) % 32) = 0) (type: boolean) + Statistics: Num rows: 250 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tst_dest30 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(a, 'hll') + mode: complete + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tst_dest30 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: a + Column Types: int + Table: default.tst_dest30 + +PREHOOK: query: insert overwrite table dest30 +select count(1) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tst_dest30 +POSTHOOK: query: insert overwrite table dest30 +select count(1) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tst_dest30 +POSTHOOK: Lineage: tst_dest30.a EXPRESSION [(src)src.null, ] +PREHOOK: query: select * from tst_dest30 +PREHOOK: type: QUERY +PREHOOK: Input: default@tst_dest30 +#### A masked pattern was here #### +POSTHOOK: query: select * from tst_dest30 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tst_dest30 +#### A masked pattern was here #### +15 +PREHOOK: query: explain +insert overwrite table dest30 +select count(1) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest30 +POSTHOOK: query: explain +insert overwrite table dest30 +select count(1) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest30 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: $f0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger($f0) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest30 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(a, 'hll') + mode: complete + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: struct) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest30 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: a + Column Types: int + Table: default.dest30 + +PREHOOK: query: insert overwrite table dest30 +select count(1) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest30 +POSTHOOK: query: insert overwrite table dest30 +select count(1) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest30 +POSTHOOK: Lineage: dest30.a EXPRESSION [] +PREHOOK: query: select * from tst_dest30 +PREHOOK: type: QUERY +PREHOOK: Input: default@tst_dest30 +#### A masked pattern was here #### +POSTHOOK: query: select * from tst_dest30 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tst_dest30 +#### A masked pattern was here #### +15 diff --git ql/src/test/results/clientpositive/llap/input32.q.out ql/src/test/results/clientpositive/llap/input32.q.out new file mode 100644 index 0000000000..690279d6f3 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input32.q.out @@ -0,0 +1,140 @@ +PREHOOK: query: create table dest32(a int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest32 +POSTHOOK: query: create table dest32(a int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest32 +PREHOOK: query: create table tst_dest32(a int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tst_dest32 +POSTHOOK: query: create table tst_dest32(a int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tst_dest32 +PREHOOK: query: explain +insert overwrite table dest32 +select count(1) from srcbucket +PREHOOK: type: QUERY +PREHOOK: Input: default@srcbucket +PREHOOK: Output: default@tst_dest32 +POSTHOOK: query: explain +insert overwrite table dest32 +select count(1) from srcbucket +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcbucket +POSTHOOK: Output: default@tst_dest32 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcbucket + Statistics: Num rows: 1000 Data size: 10603 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1000 Data size: 10603 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tst_dest32 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: a + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(a, 'hll') + mode: complete + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tst_dest32 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: a + Column Types: int + Table: default.tst_dest32 + +PREHOOK: query: insert overwrite table dest32 +select count(1) from srcbucket +PREHOOK: type: QUERY +PREHOOK: Input: default@srcbucket +PREHOOK: Output: default@tst_dest32 +POSTHOOK: query: insert overwrite table dest32 +select count(1) from srcbucket +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcbucket +POSTHOOK: Output: default@tst_dest32 +POSTHOOK: Lineage: tst_dest32.a EXPRESSION [(srcbucket)srcbucket.null, ] +PREHOOK: query: select * from tst_dest32 +PREHOOK: type: QUERY +PREHOOK: Input: default@tst_dest32 +#### A masked pattern was here #### +POSTHOOK: query: select * from tst_dest32 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tst_dest32 +#### A masked pattern was here #### +1000 diff --git ql/src/test/results/clientpositive/llap/input3_limit.q.out ql/src/test/results/clientpositive/llap/input3_limit.q.out new file mode 100644 index 0000000000..c9616de7a2 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input3_limit.q.out @@ -0,0 +1,202 @@ +PREHOOK: query: CREATE TABLE T1_n78(key STRING, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n78 +POSTHOOK: query: CREATE TABLE T1_n78(key STRING, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n78 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE T1_n78 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n78 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv1.txt' INTO TABLE T1_n78 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n78 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv2.txt' INTO TABLE T1_n78 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n78 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/kv2.txt' INTO TABLE T1_n78 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n78 +PREHOOK: query: CREATE TABLE T2_n47(key STRING, value STRING) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n47 +POSTHOOK: query: CREATE TABLE T2_n47(key STRING, value STRING) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n47 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n78 +PREHOOK: Output: default@t2_n47 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n78 +POSTHOOK: Output: default@t2_n47 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t1_n78 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Top N Key Operator + sort order: ++ + keys: VALUE._col0 (type: string), VALUE._col1 (type: string) + null sort order: zz + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + top n: 20 + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 20 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n47 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.t2_n47 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.t2_n47 + +PREHOOK: query: INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n78 +PREHOOK: Output: default@t2_n47 +POSTHOOK: query: INSERT OVERWRITE TABLE T2_n47 SELECT * FROM (SELECT * FROM T1_n78 DISTRIBUTE BY key) T ORDER BY key, value LIMIT 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n78 +POSTHOOK: Output: default@t2_n47 +POSTHOOK: Lineage: t2_n47.key SIMPLE [(t1_n78)t1_n78.FieldSchema(name:key, type:string, comment:null), ] +POSTHOOK: Lineage: t2_n47.value SIMPLE [(t1_n78)t1_n78.FieldSchema(name:value, type:string, comment:null), ] +PREHOOK: query: SELECT * FROM T2_n47 ORDER BY key, value +PREHOOK: type: QUERY +PREHOOK: Input: default@t2_n47 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM T2_n47 ORDER BY key, value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t2_n47 +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +0 val_1 +0 val_1 +1 val_2 +10 val_10 +10 val_11 +100 val_100 +100 val_100 +100 val_101 +100 val_101 +101 val_102 +102 val_103 +103 val_103 +103 val_103 +104 val_104 +104 val_104 +104 val_105 +104 val_105 diff --git ql/src/test/results/clientpositive/input4.q.out ql/src/test/results/clientpositive/llap/input4.q.out similarity index 95% rename from ql/src/test/results/clientpositive/input4.q.out rename to ql/src/test/results/clientpositive/llap/input4.q.out index 60ed774591..ba6a327e91 100644 --- ql/src/test/results/clientpositive/input4.q.out +++ ql/src/test/results/clientpositive/llap/input4.q.out @@ -53,7 +53,7 @@ SELECT Input4Alias.VALUE, Input4Alias.KEY FROM INPUT4_n0 AS Input4Alias POSTHOOK: type: QUERY POSTHOOK: Input: default@input4_n0 #### A masked pattern was here #### -{"CBOPlan":"{\n \"rels\": [\n {\n \"id\": \"0\",\n \"relOp\": \"org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveTableScan\",\n \"table\": [\n \"default\",\n \"input4_n0\"\n ],\n \"table:alias\": \"input4alias\",\n \"inputs\": [],\n \"rowCount\": 1.0,\n \"avgRowSize\": 200.0,\n \"rowType\": [\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"key\"\n },\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"value\"\n },\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"BLOCK__OFFSET__INSIDE__FILE\"\n },\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"INPUT__FILE__NAME\"\n },\n {\n \"fields\": [\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"writeid\"\n },\n {\n \"type\": \"INTEGER\",\n \"nullable\": true,\n \"name\": \"bucketid\"\n },\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"rowid\"\n }\n ],\n \"name\": \"ROW__ID\"\n }\n ],\n \"colStats\": [\n {\n \"name\": \"key\",\n \"ndv\": 1\n },\n {\n \"name\": \"value\",\n \"ndv\": 1\n }\n ]\n },\n {\n \"id\": \"1\",\n \"relOp\": \"org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveProject\",\n \"fields\": [\n \"value\",\n \"key\"\n ],\n \"exprs\": [\n {\n \"input\": 1,\n \"name\": \"$1\"\n },\n {\n \"input\": 0,\n \"name\": \"$0\"\n }\n ],\n \"rowCount\": 1.0\n }\n ]\n}","optimizedSQL":"SELECT `value`, `key`\nFROM `default`.`input4_n0`","cboInfo":"Plan optimized by CBO.","STAGE DEPENDENCIES":{"Stage-0":{"ROOT STAGE":"TRUE"}},"STAGE PLANS":{"Stage-0":{"Fetch Operator":{"limit:":"-1","Processor Tree:":{"TableScan":{"alias:":"input4alias","columns:":["value","key"],"database:":"default","Statistics:":"Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE","table:":"input4_n0","isTempTable:":"false","OperatorId:":"TS_0","children":{"Select Operator":{"expressions:":"value (type: string), key (type: string)","columnExprMap:":{"_col0":"value","_col1":"key"},"outputColumnNames:":["_col0","_col1"],"Statistics:":"Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE","OperatorId:":"SEL_1","children":{"ListSink":{"OperatorId:":"LIST_SINK_3"}}}}}}}}}} +{"CBOPlan":"{\n \"rels\": [\n {\n \"id\": \"0\",\n \"relOp\": \"org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveTableScan\",\n \"table\": [\n \"default\",\n \"input4_n0\"\n ],\n \"table:alias\": \"input4alias\",\n \"inputs\": [],\n \"rowCount\": 1.0,\n \"avgRowSize\": 200.0,\n \"rowType\": [\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"key\"\n },\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"value\"\n },\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"BLOCK__OFFSET__INSIDE__FILE\"\n },\n {\n \"type\": \"VARCHAR\",\n \"nullable\": true,\n \"precision\": 2147483647,\n \"name\": \"INPUT__FILE__NAME\"\n },\n {\n \"fields\": [\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"writeid\"\n },\n {\n \"type\": \"INTEGER\",\n \"nullable\": true,\n \"name\": \"bucketid\"\n },\n {\n \"type\": \"BIGINT\",\n \"nullable\": true,\n \"name\": \"rowid\"\n }\n ],\n \"name\": \"ROW__ID\"\n }\n ],\n \"colStats\": [\n {\n \"name\": \"key\",\n \"ndv\": 1\n },\n {\n \"name\": \"value\",\n \"ndv\": 1\n }\n ]\n },\n {\n \"id\": \"1\",\n \"relOp\": \"org.apache.hadoop.hive.ql.optimizer.calcite.reloperators.HiveProject\",\n \"fields\": [\n \"value\",\n \"key\"\n ],\n \"exprs\": [\n {\n \"input\": 1,\n \"name\": \"$1\"\n },\n {\n \"input\": 0,\n \"name\": \"$0\"\n }\n ],\n \"rowCount\": 1.0\n }\n ]\n}","optimizedSQL":"SELECT `value`, `key`\nFROM `default`.`input4_n0`","cboInfo":"Plan optimized by CBO.","STAGE DEPENDENCIES":{"Stage-0":{"ROOT STAGE":"TRUE"}},"STAGE PLANS":{"Stage-0":{"Fetch Operator":{"limit:":"-1","Processor Tree:":{"TableScan":{"alias:":"input4alias","columns:":["value","key"],"database:":"default","table:":"input4_n0","isTempTable:":"false","OperatorId:":"TS_0","children":{"Select Operator":{"expressions:":"value (type: string), key (type: string)","columnExprMap:":{"_col0":"value","_col1":"key"},"outputColumnNames:":["_col0","_col1"],"OperatorId:":"SEL_1","children":{"ListSink":{"OperatorId:":"LIST_SINK_3"}}}}}}}}}} PREHOOK: query: SELECT Input4Alias.VALUE, Input4Alias.KEY FROM INPUT4_n0 AS Input4Alias PREHOOK: type: QUERY PREHOOK: Input: default@input4_n0 diff --git ql/src/test/results/clientpositive/input42.q.out ql/src/test/results/clientpositive/llap/input42.q.out similarity index 91% rename from ql/src/test/results/clientpositive/input42.q.out rename to ql/src/test/results/clientpositive/llap/input42.q.out index 15c363bc5f..20d204cd54 100644 --- ql/src/test/results/clientpositive/input42.q.out +++ ql/src/test/results/clientpositive/llap/input42.q.out @@ -121,12 +121,10 @@ STAGE PLANS: TableScan alias: a filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE GatherStats: false Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: select * from srcpart a where a.ds='2008-04-08' @@ -1159,57 +1157,14 @@ OPTIMIZED SQL: SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`, `hr` FROM `default`.`srcpart` WHERE `key` < 200 AND `ds` = '2008-04-08' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) < 200.0D) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (UDFToDouble(key) < 200.0D) (type: boolean) - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=11 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -1256,9 +1211,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart -#### A masked pattern was here #### Partition - base file name: hr=12 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -1305,15 +1258,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [a] - /srcpart/ds=2008-04-08/hr=12 [a] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: a + filterExpr: ((UDFToDouble(key) < 200.0D) and (ds = '2008-04-08')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) < 200.0D) (type: boolean) + Select Operator + expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from srcpart a where a.ds='2008-04-08' and key < 200 PREHOOK: type: QUERY @@ -1723,56 +1679,14 @@ OPTIMIZED SQL: SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`, `hr` FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' AND RAND(100) < 0.1 STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (rand(100) < 0.1D) (type: boolean) - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=11 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -1819,9 +1733,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart -#### A masked pattern was here #### Partition - base file name: hr=12 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -1868,15 +1780,17 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [a] - /srcpart/ds=2008-04-08/hr=12 [a] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: a + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (rand(100) < 0.1D) (type: boolean) + Select Operator + expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from srcpart a where a.ds='2008-04-08' and rand(100) < 0.1 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/input4_limit.q.out ql/src/test/results/clientpositive/llap/input4_limit.q.out new file mode 100644 index 0000000000..fad0fb62cf --- /dev/null +++ ql/src/test/results/clientpositive/llap/input4_limit.q.out @@ -0,0 +1,112 @@ +PREHOOK: query: explain +select * from src sort by key limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from src sort by key limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: key (type: string) + null sort order: z + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: + + keys: _col0 (type: string) + null sort order: z + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: string) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 1780 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: select * from src sort by key limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from src sort by key limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +10 val_10 +100 val_100 +100 val_100 +103 val_103 +103 val_103 +104 val_104 +104 val_104 diff --git ql/src/test/results/clientpositive/llap/input6.q.out ql/src/test/results/clientpositive/llap/input6.q.out new file mode 100644 index 0000000000..6565abd3b9 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input6.q.out @@ -0,0 +1,129 @@ +PREHOOK: query: CREATE TABLE dest1_n35(key STRING, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n35 +POSTHOOK: query: CREATE TABLE dest1_n35(key STRING, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n35 +PREHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n35 +POSTHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n35 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is null (type: boolean) + Statistics: Num rows: 1 Data size: 175 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: null (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n35 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 1 Data size: 173 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n35 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.dest1_n35 + +PREHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n35 +POSTHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n35 SELECT src1.key, src1.value WHERE src1.key is null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n35 +POSTHOOK: Lineage: dest1_n35.key SIMPLE [] +POSTHOOK: Lineage: dest1_n35.value SIMPLE [(src1)src1.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n35.* FROM dest1_n35 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n35 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n35.* FROM dest1_n35 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n35 +#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/llap/input7.q.out ql/src/test/results/clientpositive/llap/input7.q.out new file mode 100644 index 0000000000..88afecd8c3 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input7.q.out @@ -0,0 +1,150 @@ +PREHOOK: query: CREATE TABLE dest1_n167(c1 DOUBLE, c2 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n167 +POSTHOOK: query: CREATE TABLE dest1_n167(c1 DOUBLE, c2 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n167 +PREHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n167 +POSTHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n167 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: null (type: double), UDFToInteger(key) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n167 + Select Operator + expressions: _col0 (type: double), _col1 (type: int) + outputColumnNames: c1, c2 + Statistics: Num rows: 25 Data size: 108 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') + minReductionHashAggr: 0.96 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n167 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2 + Column Types: double, int + Table: default.dest1_n167 + +PREHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n167 +POSTHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n167 SELECT NULL, src1.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n167 +POSTHOOK: Lineage: dest1_n167.c1 EXPRESSION [] +POSTHOOK: Lineage: dest1_n167.c2 EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n167.* FROM dest1_n167 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n167 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n167.* FROM dest1_n167 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n167 +#### A masked pattern was here #### +NULL 238 +NULL NULL +NULL 311 +NULL NULL +NULL NULL +NULL NULL +NULL 255 +NULL 278 +NULL 98 +NULL NULL +NULL NULL +NULL NULL +NULL 401 +NULL 150 +NULL 273 +NULL 224 +NULL 369 +NULL 66 +NULL 128 +NULL 213 +NULL 146 +NULL 406 +NULL NULL +NULL NULL +NULL NULL diff --git ql/src/test/results/clientpositive/llap/input8.q.out ql/src/test/results/clientpositive/llap/input8.q.out new file mode 100644 index 0000000000..6e8c5a1135 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input8.q.out @@ -0,0 +1,151 @@ +PREHOOK: query: CREATE TABLE dest1_n28(c1 STRING, c2 INT, c3 DOUBLE) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n28 +POSTHOOK: query: CREATE TABLE dest1_n28(c1 STRING, c2 INT, c3 DOUBLE) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n28 +PREHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n28 +POSTHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n28 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 191 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: null (type: string), null (type: int), null (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n28 + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: double) + outputColumnNames: c1, c2, c3 + Statistics: Num rows: 25 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll') + minReductionHashAggr: 0.96 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n28 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3 + Column Types: string, int, double + Table: default.dest1_n28 + +PREHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n28 +POSTHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n28 SELECT 4 + NULL, src1.key - NULL, NULL + NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n28 +POSTHOOK: Lineage: dest1_n28.c1 EXPRESSION [] +POSTHOOK: Lineage: dest1_n28.c2 EXPRESSION [] +POSTHOOK: Lineage: dest1_n28.c3 EXPRESSION [] +PREHOOK: query: SELECT dest1_n28.* FROM dest1_n28 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n28 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n28.* FROM dest1_n28 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n28 +#### A masked pattern was here #### +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL +NULL NULL NULL diff --git ql/src/test/results/clientpositive/llap/input9.q.out ql/src/test/results/clientpositive/llap/input9.q.out new file mode 100644 index 0000000000..dda00b30e1 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input9.q.out @@ -0,0 +1,132 @@ +PREHOOK: query: CREATE TABLE dest1_n159(value STRING, key INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n159 +POSTHOOK: query: CREATE TABLE dest1_n159(value STRING, key INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n159 +PREHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n159 +POSTHOOK: query: EXPLAIN +FROM src1 +INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n159 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 86 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: null (type: string), UDFToInteger(_col0) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n159 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: value, key + Statistics: Num rows: 1 Data size: 88 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(value, 'hll'), compute_stats(key, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n159 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: value, key + Column Types: string, int + Table: default.dest1_n159 + +PREHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest1_n159 +POSTHOOK: query: FROM src1 +INSERT OVERWRITE TABLE dest1_n159 SELECT NULL, src1.key where NULL = NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest1_n159 +POSTHOOK: Lineage: dest1_n159.key EXPRESSION [(src1)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n159.value EXPRESSION [] +PREHOOK: query: SELECT dest1_n159.* FROM dest1_n159 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n159 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n159.* FROM dest1_n159 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n159 +#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/input_columnarserde.q.out ql/src/test/results/clientpositive/llap/input_columnarserde.q.out similarity index 69% rename from ql/src/test/results/clientpositive/input_columnarserde.q.out rename to ql/src/test/results/clientpositive/llap/input_columnarserde.q.out index ff065730c4..57c8c8eccc 100644 --- ql/src/test/results/clientpositive/input_columnarserde.q.out +++ ql/src/test/results/clientpositive/llap/input_columnarserde.q.out @@ -30,40 +30,53 @@ POSTHOOK: Input: default@src_thrift POSTHOOK: Output: default@input_columnarserde STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Map-reduce partition columns: 1 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: 1 (type: int) + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.input_columnarserde + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.input_columnarserde + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -75,7 +88,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.input_columnarserde - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/input_dynamicserde.q.out ql/src/test/results/clientpositive/llap/input_dynamicserde.q.out similarity index 66% rename from ql/src/test/results/clientpositive/input_dynamicserde.q.out rename to ql/src/test/results/clientpositive/llap/input_dynamicserde.q.out index 226bad75ee..f88cab5213 100644 --- ql/src/test/results/clientpositive/input_dynamicserde.q.out +++ ql/src/test/results/clientpositive/llap/input_dynamicserde.q.out @@ -32,43 +32,37 @@ POSTHOOK: Input: default@src_thrift POSTHOOK: Output: default@dest1_n114 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n114 - Execution mode: vectorized - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Tez #### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n114 + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -80,40 +74,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n114 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n114 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n114 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: FROM src_thrift INSERT OVERWRITE TABLE dest1_n114 SELECT src_thrift.lint, src_thrift.lstring, src_thrift.mstringstring, src_thrift.aint, src_thrift.astring PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/input_lazyserde.q.out ql/src/test/results/clientpositive/llap/input_lazyserde.q.out similarity index 90% rename from ql/src/test/results/clientpositive/input_lazyserde.q.out rename to ql/src/test/results/clientpositive/llap/input_lazyserde.q.out index 79cc9e72dc..016dfba53b 100644 --- ql/src/test/results/clientpositive/input_lazyserde.q.out +++ ql/src/test/results/clientpositive/llap/input_lazyserde.q.out @@ -36,39 +36,53 @@ POSTHOOK: Input: default@src_thrift POSTHOOK: Output: default@dest1_n43 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Map-reduce partition columns: 1 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: 1 (type: int) + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n43 + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n43 + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -80,7 +94,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n43 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/input_lazyserde2.q.out ql/src/test/results/clientpositive/llap/input_lazyserde2.q.out similarity index 90% rename from ql/src/test/results/clientpositive/input_lazyserde2.q.out rename to ql/src/test/results/clientpositive/llap/input_lazyserde2.q.out index 5f3adcfb8a..a831c6a4c4 100644 --- ql/src/test/results/clientpositive/input_lazyserde2.q.out +++ ql/src/test/results/clientpositive/llap/input_lazyserde2.q.out @@ -36,39 +36,53 @@ POSTHOOK: Input: default@src_thrift POSTHOOK: Output: default@dest1_n148 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Map-reduce partition columns: 1 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: lint (type: array), lstring (type: array), mstringstring (type: map), aint (type: int), astring (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Map-reduce partition columns: 1 (type: int) + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: array), _col1 (type: array), _col2 (type: map), _col3 (type: int), _col4 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: array), VALUE._col1 (type: array), VALUE._col2 (type: map), VALUE._col3 (type: int), VALUE._col4 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n148 + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 34628 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n148 + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -80,7 +94,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n148 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/input_limit.q.out ql/src/test/results/clientpositive/llap/input_limit.q.out similarity index 79% rename from ql/src/test/results/clientpositive/input_limit.q.out rename to ql/src/test/results/clientpositive/llap/input_limit.q.out index 2cbcc36a0a..78eedf0726 100644 --- ql/src/test/results/clientpositive/input_limit.q.out +++ ql/src/test/results/clientpositive/llap/input_limit.q.out @@ -18,14 +18,11 @@ STAGE PLANS: Processor Tree: TableScan alias: x - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 20 - Statistics: Num rows: 20 Data size: 3560 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRC x LIMIT 20 diff --git ql/src/test/results/clientpositive/input_part0.q.out ql/src/test/results/clientpositive/llap/input_part0.q.out similarity index 99% rename from ql/src/test/results/clientpositive/input_part0.q.out rename to ql/src/test/results/clientpositive/llap/input_part0.q.out index 7d02f9fac8..3466f5a8a6 100644 --- ql/src/test/results/clientpositive/input_part0.q.out +++ ql/src/test/results/clientpositive/llap/input_part0.q.out @@ -23,11 +23,9 @@ STAGE PLANS: TableScan alias: x filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' diff --git ql/src/test/results/clientpositive/llap/input_part1.q.out ql/src/test/results/clientpositive/llap/input_part1.q.out new file mode 100644 index 0000000000..eb74be644c --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_part1.q.out @@ -0,0 +1,347 @@ +PREHOOK: query: CREATE TABLE dest1_n45(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n45 +POSTHOOK: query: CREATE TABLE dest1_n45(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n45 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart +INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n45 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart +INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n45 +OPTIMIZED SQL: SELECT `key`, `value`, CAST('12' AS STRING) AS `hr`, CAST('2008-04-08' AS STRING) AS `ds` +FROM `default`.`srcpart` +WHERE `key` < 100 AND `ds` = '2008-04-08' AND `hr` = '12' +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08') and (hr = '12')) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest1_n45 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n45 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, hr, ds + Statistics: Num rows: 166 Data size: 45650 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest1_n45 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n45 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, hr, ds + Column Types: int, string, string, string + Table: default.dest1_n45 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart +INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@dest1_n45 +POSTHOOK: query: FROM srcpart +INSERT OVERWRITE TABLE dest1_n45 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@dest1_n45 +POSTHOOK: Lineage: dest1_n45.ds SIMPLE [] +POSTHOOK: Lineage: dest1_n45.hr SIMPLE [] +POSTHOOK: Lineage: dest1_n45.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n45.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n45.* FROM dest1_n45 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n45 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n45.* FROM dest1_n45 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n45 +#### A masked pattern was here #### +86 val_86 12 2008-04-08 +27 val_27 12 2008-04-08 +98 val_98 12 2008-04-08 +66 val_66 12 2008-04-08 +37 val_37 12 2008-04-08 +15 val_15 12 2008-04-08 +82 val_82 12 2008-04-08 +17 val_17 12 2008-04-08 +0 val_0 12 2008-04-08 +57 val_57 12 2008-04-08 +20 val_20 12 2008-04-08 +92 val_92 12 2008-04-08 +47 val_47 12 2008-04-08 +72 val_72 12 2008-04-08 +4 val_4 12 2008-04-08 +35 val_35 12 2008-04-08 +54 val_54 12 2008-04-08 +51 val_51 12 2008-04-08 +65 val_65 12 2008-04-08 +83 val_83 12 2008-04-08 +12 val_12 12 2008-04-08 +67 val_67 12 2008-04-08 +84 val_84 12 2008-04-08 +58 val_58 12 2008-04-08 +8 val_8 12 2008-04-08 +24 val_24 12 2008-04-08 +42 val_42 12 2008-04-08 +0 val_0 12 2008-04-08 +96 val_96 12 2008-04-08 +26 val_26 12 2008-04-08 +51 val_51 12 2008-04-08 +43 val_43 12 2008-04-08 +95 val_95 12 2008-04-08 +98 val_98 12 2008-04-08 +85 val_85 12 2008-04-08 +77 val_77 12 2008-04-08 +0 val_0 12 2008-04-08 +87 val_87 12 2008-04-08 +15 val_15 12 2008-04-08 +72 val_72 12 2008-04-08 +90 val_90 12 2008-04-08 +19 val_19 12 2008-04-08 +10 val_10 12 2008-04-08 +5 val_5 12 2008-04-08 +58 val_58 12 2008-04-08 +35 val_35 12 2008-04-08 +95 val_95 12 2008-04-08 +11 val_11 12 2008-04-08 +34 val_34 12 2008-04-08 +42 val_42 12 2008-04-08 +78 val_78 12 2008-04-08 +76 val_76 12 2008-04-08 +41 val_41 12 2008-04-08 +30 val_30 12 2008-04-08 +64 val_64 12 2008-04-08 +76 val_76 12 2008-04-08 +74 val_74 12 2008-04-08 +69 val_69 12 2008-04-08 +33 val_33 12 2008-04-08 +70 val_70 12 2008-04-08 +5 val_5 12 2008-04-08 +2 val_2 12 2008-04-08 +35 val_35 12 2008-04-08 +80 val_80 12 2008-04-08 +44 val_44 12 2008-04-08 +53 val_53 12 2008-04-08 +90 val_90 12 2008-04-08 +12 val_12 12 2008-04-08 +5 val_5 12 2008-04-08 +70 val_70 12 2008-04-08 +24 val_24 12 2008-04-08 +70 val_70 12 2008-04-08 +83 val_83 12 2008-04-08 +26 val_26 12 2008-04-08 +67 val_67 12 2008-04-08 +18 val_18 12 2008-04-08 +9 val_9 12 2008-04-08 +18 val_18 12 2008-04-08 +97 val_97 12 2008-04-08 +84 val_84 12 2008-04-08 +28 val_28 12 2008-04-08 +37 val_37 12 2008-04-08 +90 val_90 12 2008-04-08 +97 val_97 12 2008-04-08 diff --git ql/src/test/results/clientpositive/llap/input_part10.q.out ql/src/test/results/clientpositive/llap/input_part10.q.out new file mode 100644 index 0000000000..dac65c50a4 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_part10.q.out @@ -0,0 +1,192 @@ +PREHOOK: query: CREATE TABLE part_special ( + a STRING, + b STRING +) PARTITIONED BY ( + ds STRING, + ts STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part_special +POSTHOOK: query: CREATE TABLE part_special ( + a STRING, + b STRING +) PARTITIONED BY ( + ds STRING, + ts STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part_special +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +SELECT 1, 2 FROM src LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +SELECT 1, 2 FROM src LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 1 (type: int), 2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: int), _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: CAST( _col0 AS STRING) (type: string), CAST( _col1 AS STRING) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.part_special + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008 04 08' (type: string), '10:11:12=455' (type: string) + outputColumnNames: a, b, ds, ts + Statistics: Num rows: 1 Data size: 558 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(a, 'hll'), compute_stats(b, 'hll') + keys: ds (type: string), ts (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1070 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008 04 08 + ts 10:11:12=455 + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.part_special + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: a, b + Column Types: string, string + Table: default.part_special + +PREHOOK: query: INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +SELECT 1, 2 FROM src LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +POSTHOOK: query: INSERT OVERWRITE TABLE part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +SELECT 1, 2 FROM src LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).a EXPRESSION [] +POSTHOOK: Lineage: part_special PARTITION(ds=2008 04 08,ts=10:11:12=455).b EXPRESSION [] +PREHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@part_special +POSTHOOK: query: DESCRIBE EXTENDED part_special PARTITION(ds='2008 04 08', ts = '10:11:12=455') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@part_special +a string +b string +ds string +ts string + +# Partition Information +# col_name data_type comment +ds string +ts string + +#### A masked pattern was here #### +PREHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' +PREHOOK: type: QUERY +PREHOOK: Input: default@part_special +PREHOOK: Input: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM part_special WHERE ds='2008 04 08' AND ts = '10:11:12=455' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part_special +POSTHOOK: Input: default@part_special@ds=2008 04 08/ts=10%3A11%3A12%3D455 +#### A masked pattern was here #### +1 2 2008 04 08 10:11:12=455 diff --git ql/src/test/results/clientpositive/llap/input_part2.q.out ql/src/test/results/clientpositive/llap/input_part2.q.out new file mode 100644 index 0000000000..3a5dde72f8 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_part2.q.out @@ -0,0 +1,645 @@ +PREHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n84 +POSTHOOK: query: CREATE TABLE dest1_n84(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n84 +PREHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest2_n20 +POSTHOOK: query: CREATE TABLE dest2_n20(key INT, value STRING, hr STRING, ds STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest2_n20 +PREHOOK: query: EXPLAIN EXTENDED +FROM srcpart +INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +PREHOOK: Output: default@dest1_n84 +PREHOOK: Output: default@dest2_n20 +POSTHOOK: query: EXPLAIN EXTENDED +FROM srcpart +INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +POSTHOOK: Output: default@dest1_n84 +POSTHOOK: Output: default@dest2_n20 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + Statistics: Num rows: 1000 Data size: 272000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key < 100) and (ds = '2008-04-08')) (type: boolean) + Statistics: Num rows: 333 Data size: 90576 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-08' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest1_n84 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n84 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-08' (type: string) + outputColumnNames: key, value, hr, ds + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Filter Operator + isSamplingPred: false + predicate: ((key < 100) and (ds = '2008-04-09')) (type: boolean) + Statistics: Num rows: 333 Data size: 90576 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string), '12' (type: string), '2008-04-09' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 2 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest2_n20 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n20 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), '12' (type: string), '2008-04-09' (type: string) + outputColumnNames: key, value, hr, ds + Statistics: Num rows: 333 Data size: 91575 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1744 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=12 [srcpart] + /srcpart/ds=2008-04-09/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest1_n84 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n84 + + Stage: Stage-4 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, hr, ds + Column Types: int, string, string, string + Table: default.dest1_n84 + Is Table Level Stats: true + + Stage: Stage-1 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,hr,ds + columns.comments + columns.types int:string:string:string +#### A masked pattern was here #### + name default.dest2_n20 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest2_n20 + + Stage: Stage-5 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, hr, ds + Column Types: int, string, string, string + Table: default.dest2_n20 + Is Table Level Stats: true + +PREHOOK: query: FROM srcpart +INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +PREHOOK: Output: default@dest1_n84 +PREHOOK: Output: default@dest2_n20 +POSTHOOK: query: FROM srcpart +INSERT OVERWRITE TABLE dest1_n84 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-08' and srcpart.hr = '12' +INSERT OVERWRITE TABLE dest2_n20 SELECT srcpart.key, srcpart.value, srcpart.hr, srcpart.ds WHERE srcpart.key < 100 and srcpart.ds = '2008-04-09' and srcpart.hr = '12' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +POSTHOOK: Output: default@dest1_n84 +POSTHOOK: Output: default@dest2_n20 +POSTHOOK: Lineage: dest1_n84.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] +POSTHOOK: Lineage: dest1_n84.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] +POSTHOOK: Lineage: dest1_n84.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n84.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n20.ds SIMPLE [(srcpart)srcpart.FieldSchema(name:ds, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n20.hr SIMPLE [(srcpart)srcpart.FieldSchema(name:hr, type:string, comment:null), ] +POSTHOOK: Lineage: dest2_n20.key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest2_n20.value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n84 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n84.* FROM dest1_n84 sort by key,value,ds,hr +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n84 +#### A masked pattern was here #### +0 val_0 12 2008-04-08 +0 val_0 12 2008-04-08 +0 val_0 12 2008-04-08 +10 val_10 12 2008-04-08 +11 val_11 12 2008-04-08 +12 val_12 12 2008-04-08 +12 val_12 12 2008-04-08 +15 val_15 12 2008-04-08 +15 val_15 12 2008-04-08 +17 val_17 12 2008-04-08 +18 val_18 12 2008-04-08 +18 val_18 12 2008-04-08 +19 val_19 12 2008-04-08 +2 val_2 12 2008-04-08 +20 val_20 12 2008-04-08 +24 val_24 12 2008-04-08 +24 val_24 12 2008-04-08 +26 val_26 12 2008-04-08 +26 val_26 12 2008-04-08 +27 val_27 12 2008-04-08 +28 val_28 12 2008-04-08 +30 val_30 12 2008-04-08 +33 val_33 12 2008-04-08 +34 val_34 12 2008-04-08 +35 val_35 12 2008-04-08 +35 val_35 12 2008-04-08 +35 val_35 12 2008-04-08 +37 val_37 12 2008-04-08 +37 val_37 12 2008-04-08 +4 val_4 12 2008-04-08 +41 val_41 12 2008-04-08 +42 val_42 12 2008-04-08 +42 val_42 12 2008-04-08 +43 val_43 12 2008-04-08 +44 val_44 12 2008-04-08 +47 val_47 12 2008-04-08 +5 val_5 12 2008-04-08 +5 val_5 12 2008-04-08 +5 val_5 12 2008-04-08 +51 val_51 12 2008-04-08 +51 val_51 12 2008-04-08 +53 val_53 12 2008-04-08 +54 val_54 12 2008-04-08 +57 val_57 12 2008-04-08 +58 val_58 12 2008-04-08 +58 val_58 12 2008-04-08 +64 val_64 12 2008-04-08 +65 val_65 12 2008-04-08 +66 val_66 12 2008-04-08 +67 val_67 12 2008-04-08 +67 val_67 12 2008-04-08 +69 val_69 12 2008-04-08 +70 val_70 12 2008-04-08 +70 val_70 12 2008-04-08 +70 val_70 12 2008-04-08 +72 val_72 12 2008-04-08 +72 val_72 12 2008-04-08 +74 val_74 12 2008-04-08 +76 val_76 12 2008-04-08 +76 val_76 12 2008-04-08 +77 val_77 12 2008-04-08 +78 val_78 12 2008-04-08 +8 val_8 12 2008-04-08 +80 val_80 12 2008-04-08 +82 val_82 12 2008-04-08 +83 val_83 12 2008-04-08 +83 val_83 12 2008-04-08 +84 val_84 12 2008-04-08 +84 val_84 12 2008-04-08 +85 val_85 12 2008-04-08 +86 val_86 12 2008-04-08 +87 val_87 12 2008-04-08 +9 val_9 12 2008-04-08 +90 val_90 12 2008-04-08 +90 val_90 12 2008-04-08 +90 val_90 12 2008-04-08 +92 val_92 12 2008-04-08 +95 val_95 12 2008-04-08 +95 val_95 12 2008-04-08 +96 val_96 12 2008-04-08 +97 val_97 12 2008-04-08 +97 val_97 12 2008-04-08 +98 val_98 12 2008-04-08 +98 val_98 12 2008-04-08 +PREHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr +PREHOOK: type: QUERY +PREHOOK: Input: default@dest2_n20 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest2_n20.* FROM dest2_n20 sort by key,value,ds,hr +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest2_n20 +#### A masked pattern was here #### +0 val_0 12 2008-04-09 +0 val_0 12 2008-04-09 +0 val_0 12 2008-04-09 +10 val_10 12 2008-04-09 +11 val_11 12 2008-04-09 +12 val_12 12 2008-04-09 +12 val_12 12 2008-04-09 +15 val_15 12 2008-04-09 +15 val_15 12 2008-04-09 +17 val_17 12 2008-04-09 +18 val_18 12 2008-04-09 +18 val_18 12 2008-04-09 +19 val_19 12 2008-04-09 +2 val_2 12 2008-04-09 +20 val_20 12 2008-04-09 +24 val_24 12 2008-04-09 +24 val_24 12 2008-04-09 +26 val_26 12 2008-04-09 +26 val_26 12 2008-04-09 +27 val_27 12 2008-04-09 +28 val_28 12 2008-04-09 +30 val_30 12 2008-04-09 +33 val_33 12 2008-04-09 +34 val_34 12 2008-04-09 +35 val_35 12 2008-04-09 +35 val_35 12 2008-04-09 +35 val_35 12 2008-04-09 +37 val_37 12 2008-04-09 +37 val_37 12 2008-04-09 +4 val_4 12 2008-04-09 +41 val_41 12 2008-04-09 +42 val_42 12 2008-04-09 +42 val_42 12 2008-04-09 +43 val_43 12 2008-04-09 +44 val_44 12 2008-04-09 +47 val_47 12 2008-04-09 +5 val_5 12 2008-04-09 +5 val_5 12 2008-04-09 +5 val_5 12 2008-04-09 +51 val_51 12 2008-04-09 +51 val_51 12 2008-04-09 +53 val_53 12 2008-04-09 +54 val_54 12 2008-04-09 +57 val_57 12 2008-04-09 +58 val_58 12 2008-04-09 +58 val_58 12 2008-04-09 +64 val_64 12 2008-04-09 +65 val_65 12 2008-04-09 +66 val_66 12 2008-04-09 +67 val_67 12 2008-04-09 +67 val_67 12 2008-04-09 +69 val_69 12 2008-04-09 +70 val_70 12 2008-04-09 +70 val_70 12 2008-04-09 +70 val_70 12 2008-04-09 +72 val_72 12 2008-04-09 +72 val_72 12 2008-04-09 +74 val_74 12 2008-04-09 +76 val_76 12 2008-04-09 +76 val_76 12 2008-04-09 +77 val_77 12 2008-04-09 +78 val_78 12 2008-04-09 +8 val_8 12 2008-04-09 +80 val_80 12 2008-04-09 +82 val_82 12 2008-04-09 +83 val_83 12 2008-04-09 +83 val_83 12 2008-04-09 +84 val_84 12 2008-04-09 +84 val_84 12 2008-04-09 +85 val_85 12 2008-04-09 +86 val_86 12 2008-04-09 +87 val_87 12 2008-04-09 +9 val_9 12 2008-04-09 +90 val_90 12 2008-04-09 +90 val_90 12 2008-04-09 +90 val_90 12 2008-04-09 +92 val_92 12 2008-04-09 +95 val_95 12 2008-04-09 +95 val_95 12 2008-04-09 +96 val_96 12 2008-04-09 +97 val_97 12 2008-04-09 +97 val_97 12 2008-04-09 +98 val_98 12 2008-04-09 +98 val_98 12 2008-04-09 diff --git ql/src/test/results/clientpositive/input_part3.q.out ql/src/test/results/clientpositive/llap/input_part3.q.out similarity index 98% rename from ql/src/test/results/clientpositive/input_part3.q.out rename to ql/src/test/results/clientpositive/llap/input_part3.q.out index 39a535af62..a5be3e625f 100644 --- ql/src/test/results/clientpositive/input_part3.q.out +++ ql/src/test/results/clientpositive/llap/input_part3.q.out @@ -21,11 +21,9 @@ STAGE PLANS: TableScan alias: x filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 181000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 228000 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 11 diff --git ql/src/test/results/clientpositive/input_part4.q.out ql/src/test/results/clientpositive/llap/input_part4.q.out similarity index 82% rename from ql/src/test/results/clientpositive/input_part4.q.out rename to ql/src/test/results/clientpositive/llap/input_part4.q.out index 507d901237..195fc25a7d 100644 --- ql/src/test/results/clientpositive/input_part4.q.out +++ ql/src/test/results/clientpositive/llap/input_part4.q.out @@ -19,14 +19,11 @@ STAGE PLANS: TableScan alias: x filterExpr: ((ds = '2008-04-08') and (UDFToDouble(hr) = 15.0D)) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE Filter Operator predicate: ((ds = '2008-04-08') and (UDFToDouble(hr) = 15.0D)) (type: boolean) - Statistics: Num rows: 1 Data size: 536 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 446 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.hr = 15 diff --git ql/src/test/results/clientpositive/input_part5.q.out ql/src/test/results/clientpositive/llap/input_part5.q.out similarity index 62% rename from ql/src/test/results/clientpositive/input_part5.q.out rename to ql/src/test/results/clientpositive/llap/input_part5.q.out index a7a305a8d3..cf429ac3b8 100644 --- ql/src/test/results/clientpositive/input_part5.q.out +++ ql/src/test/results/clientpositive/llap/input_part5.q.out @@ -24,74 +24,74 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Output: default@tmptable_n2 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 100.0D) (type: boolean) - Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tmptable_n2 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, hr, ds + Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable_n2 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, hr, ds - Statistics: Num rows: 333 Data size: 151848 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(hr, 'hll'), compute_stats(ds, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -103,7 +103,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tmptable_n2 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -111,36 +111,6 @@ STAGE PLANS: Column Types: string, string, string, string Table: default.tmptable_n2 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable_n2 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable_n2 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table tmptable_n2 SELECT x.* FROM SRCPART x WHERE x.ds = '2008-04-08' and x.key < 100 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/input_part6.q.out ql/src/test/results/clientpositive/llap/input_part6.q.out similarity index 86% rename from ql/src/test/results/clientpositive/input_part6.q.out rename to ql/src/test/results/clientpositive/llap/input_part6.q.out index aab21401b3..0f8c826817 100644 --- ql/src/test/results/clientpositive/input_part6.q.out +++ ql/src/test/results/clientpositive/llap/input_part6.q.out @@ -23,14 +23,11 @@ STAGE PLANS: TableScan alias: x filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 4560 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE x.ds = "2008-04-08" LIMIT 10 diff --git ql/src/test/results/clientpositive/llap/input_part7.q.out ql/src/test/results/clientpositive/llap/input_part7.q.out new file mode 100644 index 0000000000..6734bf1554 --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_part7.q.out @@ -0,0 +1,705 @@ +PREHOOK: query: EXPLAIN EXTENDED +SELECT * FROM ( + SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 + UNION ALL + SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 +) A +SORT BY A.key, A.value, A.ds, A.hr +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED +SELECT * FROM ( + SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 + UNION ALL + SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 +) A +SORT BY A.key, A.value, A.ds, A.hr +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) + null sort order: zzz + numBuckets: -1 + sort order: +++ + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [x] + /srcpart/ds=2008-04-08/hr=12 [x] + Map 4 + Map Operator Tree: + TableScan + alias: y + filterExpr: ((UDFToDouble(key) < 100.0D) and (ds = '2008-04-08')) (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 333 Data size: 120546 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) + null sort order: zzz + numBuckets: -1 + sort order: +++ + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [y] + /srcpart/ds=2008-04-08/hr=12 [y] + Reducer 3 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), '2008-04-08' (type: string), KEY.reducesinkkey2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Union 2 + Vertex: Union 2 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM ( + SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 + UNION ALL + SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 +) A +SORT BY A.key, A.value, A.ds, A.hr +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM ( + SELECT X.* FROM SRCPART X WHERE X.ds = '2008-04-08' and X.key < 100 + UNION ALL + SELECT Y.* FROM SRCPART Y WHERE Y.ds = '2008-04-08' and Y.key < 100 +) A +SORT BY A.key, A.value, A.ds, A.hr +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 11 +0 val_0 2008-04-08 12 +0 val_0 2008-04-08 12 +0 val_0 2008-04-08 12 +0 val_0 2008-04-08 12 +0 val_0 2008-04-08 12 +0 val_0 2008-04-08 12 +10 val_10 2008-04-08 11 +10 val_10 2008-04-08 11 +10 val_10 2008-04-08 12 +10 val_10 2008-04-08 12 +11 val_11 2008-04-08 11 +11 val_11 2008-04-08 11 +11 val_11 2008-04-08 12 +11 val_11 2008-04-08 12 +12 val_12 2008-04-08 11 +12 val_12 2008-04-08 11 +12 val_12 2008-04-08 11 +12 val_12 2008-04-08 11 +12 val_12 2008-04-08 12 +12 val_12 2008-04-08 12 +12 val_12 2008-04-08 12 +12 val_12 2008-04-08 12 +15 val_15 2008-04-08 11 +15 val_15 2008-04-08 11 +15 val_15 2008-04-08 11 +15 val_15 2008-04-08 11 +15 val_15 2008-04-08 12 +15 val_15 2008-04-08 12 +15 val_15 2008-04-08 12 +15 val_15 2008-04-08 12 +17 val_17 2008-04-08 11 +17 val_17 2008-04-08 11 +17 val_17 2008-04-08 12 +17 val_17 2008-04-08 12 +18 val_18 2008-04-08 11 +18 val_18 2008-04-08 11 +18 val_18 2008-04-08 11 +18 val_18 2008-04-08 11 +18 val_18 2008-04-08 12 +18 val_18 2008-04-08 12 +18 val_18 2008-04-08 12 +18 val_18 2008-04-08 12 +19 val_19 2008-04-08 11 +19 val_19 2008-04-08 11 +19 val_19 2008-04-08 12 +19 val_19 2008-04-08 12 +2 val_2 2008-04-08 11 +2 val_2 2008-04-08 11 +2 val_2 2008-04-08 12 +2 val_2 2008-04-08 12 +20 val_20 2008-04-08 11 +20 val_20 2008-04-08 11 +20 val_20 2008-04-08 12 +20 val_20 2008-04-08 12 +24 val_24 2008-04-08 11 +24 val_24 2008-04-08 11 +24 val_24 2008-04-08 11 +24 val_24 2008-04-08 11 +24 val_24 2008-04-08 12 +24 val_24 2008-04-08 12 +24 val_24 2008-04-08 12 +24 val_24 2008-04-08 12 +26 val_26 2008-04-08 11 +26 val_26 2008-04-08 11 +26 val_26 2008-04-08 11 +26 val_26 2008-04-08 11 +26 val_26 2008-04-08 12 +26 val_26 2008-04-08 12 +26 val_26 2008-04-08 12 +26 val_26 2008-04-08 12 +27 val_27 2008-04-08 11 +27 val_27 2008-04-08 11 +27 val_27 2008-04-08 12 +27 val_27 2008-04-08 12 +28 val_28 2008-04-08 11 +28 val_28 2008-04-08 11 +28 val_28 2008-04-08 12 +28 val_28 2008-04-08 12 +30 val_30 2008-04-08 11 +30 val_30 2008-04-08 11 +30 val_30 2008-04-08 12 +30 val_30 2008-04-08 12 +33 val_33 2008-04-08 11 +33 val_33 2008-04-08 11 +33 val_33 2008-04-08 12 +33 val_33 2008-04-08 12 +34 val_34 2008-04-08 11 +34 val_34 2008-04-08 11 +34 val_34 2008-04-08 12 +34 val_34 2008-04-08 12 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 11 +35 val_35 2008-04-08 12 +35 val_35 2008-04-08 12 +35 val_35 2008-04-08 12 +35 val_35 2008-04-08 12 +35 val_35 2008-04-08 12 +35 val_35 2008-04-08 12 +37 val_37 2008-04-08 11 +37 val_37 2008-04-08 11 +37 val_37 2008-04-08 11 +37 val_37 2008-04-08 11 +37 val_37 2008-04-08 12 +37 val_37 2008-04-08 12 +37 val_37 2008-04-08 12 +37 val_37 2008-04-08 12 +4 val_4 2008-04-08 11 +4 val_4 2008-04-08 11 +4 val_4 2008-04-08 12 +4 val_4 2008-04-08 12 +41 val_41 2008-04-08 11 +41 val_41 2008-04-08 11 +41 val_41 2008-04-08 12 +41 val_41 2008-04-08 12 +42 val_42 2008-04-08 11 +42 val_42 2008-04-08 11 +42 val_42 2008-04-08 11 +42 val_42 2008-04-08 11 +42 val_42 2008-04-08 12 +42 val_42 2008-04-08 12 +42 val_42 2008-04-08 12 +42 val_42 2008-04-08 12 +43 val_43 2008-04-08 11 +43 val_43 2008-04-08 11 +43 val_43 2008-04-08 12 +43 val_43 2008-04-08 12 +44 val_44 2008-04-08 11 +44 val_44 2008-04-08 11 +44 val_44 2008-04-08 12 +44 val_44 2008-04-08 12 +47 val_47 2008-04-08 11 +47 val_47 2008-04-08 11 +47 val_47 2008-04-08 12 +47 val_47 2008-04-08 12 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 11 +5 val_5 2008-04-08 12 +5 val_5 2008-04-08 12 +5 val_5 2008-04-08 12 +5 val_5 2008-04-08 12 +5 val_5 2008-04-08 12 +5 val_5 2008-04-08 12 +51 val_51 2008-04-08 11 +51 val_51 2008-04-08 11 +51 val_51 2008-04-08 11 +51 val_51 2008-04-08 11 +51 val_51 2008-04-08 12 +51 val_51 2008-04-08 12 +51 val_51 2008-04-08 12 +51 val_51 2008-04-08 12 +53 val_53 2008-04-08 11 +53 val_53 2008-04-08 11 +53 val_53 2008-04-08 12 +53 val_53 2008-04-08 12 +54 val_54 2008-04-08 11 +54 val_54 2008-04-08 11 +54 val_54 2008-04-08 12 +54 val_54 2008-04-08 12 +57 val_57 2008-04-08 11 +57 val_57 2008-04-08 11 +57 val_57 2008-04-08 12 +57 val_57 2008-04-08 12 +58 val_58 2008-04-08 11 +58 val_58 2008-04-08 11 +58 val_58 2008-04-08 11 +58 val_58 2008-04-08 11 +58 val_58 2008-04-08 12 +58 val_58 2008-04-08 12 +58 val_58 2008-04-08 12 +58 val_58 2008-04-08 12 +64 val_64 2008-04-08 11 +64 val_64 2008-04-08 11 +64 val_64 2008-04-08 12 +64 val_64 2008-04-08 12 +65 val_65 2008-04-08 11 +65 val_65 2008-04-08 11 +65 val_65 2008-04-08 12 +65 val_65 2008-04-08 12 +66 val_66 2008-04-08 11 +66 val_66 2008-04-08 11 +66 val_66 2008-04-08 12 +66 val_66 2008-04-08 12 +67 val_67 2008-04-08 11 +67 val_67 2008-04-08 11 +67 val_67 2008-04-08 11 +67 val_67 2008-04-08 11 +67 val_67 2008-04-08 12 +67 val_67 2008-04-08 12 +67 val_67 2008-04-08 12 +67 val_67 2008-04-08 12 +69 val_69 2008-04-08 11 +69 val_69 2008-04-08 11 +69 val_69 2008-04-08 12 +69 val_69 2008-04-08 12 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 11 +70 val_70 2008-04-08 12 +70 val_70 2008-04-08 12 +70 val_70 2008-04-08 12 +70 val_70 2008-04-08 12 +70 val_70 2008-04-08 12 +70 val_70 2008-04-08 12 +72 val_72 2008-04-08 11 +72 val_72 2008-04-08 11 +72 val_72 2008-04-08 11 +72 val_72 2008-04-08 11 +72 val_72 2008-04-08 12 +72 val_72 2008-04-08 12 +72 val_72 2008-04-08 12 +72 val_72 2008-04-08 12 +74 val_74 2008-04-08 11 +74 val_74 2008-04-08 11 +74 val_74 2008-04-08 12 +74 val_74 2008-04-08 12 +76 val_76 2008-04-08 11 +76 val_76 2008-04-08 11 +76 val_76 2008-04-08 11 +76 val_76 2008-04-08 11 +76 val_76 2008-04-08 12 +76 val_76 2008-04-08 12 +76 val_76 2008-04-08 12 +76 val_76 2008-04-08 12 +77 val_77 2008-04-08 11 +77 val_77 2008-04-08 11 +77 val_77 2008-04-08 12 +77 val_77 2008-04-08 12 +78 val_78 2008-04-08 11 +78 val_78 2008-04-08 11 +78 val_78 2008-04-08 12 +78 val_78 2008-04-08 12 +8 val_8 2008-04-08 11 +8 val_8 2008-04-08 11 +8 val_8 2008-04-08 12 +8 val_8 2008-04-08 12 +80 val_80 2008-04-08 11 +80 val_80 2008-04-08 11 +80 val_80 2008-04-08 12 +80 val_80 2008-04-08 12 +82 val_82 2008-04-08 11 +82 val_82 2008-04-08 11 +82 val_82 2008-04-08 12 +82 val_82 2008-04-08 12 +83 val_83 2008-04-08 11 +83 val_83 2008-04-08 11 +83 val_83 2008-04-08 11 +83 val_83 2008-04-08 11 +83 val_83 2008-04-08 12 +83 val_83 2008-04-08 12 +83 val_83 2008-04-08 12 +83 val_83 2008-04-08 12 +84 val_84 2008-04-08 11 +84 val_84 2008-04-08 11 +84 val_84 2008-04-08 11 +84 val_84 2008-04-08 11 +84 val_84 2008-04-08 12 +84 val_84 2008-04-08 12 +84 val_84 2008-04-08 12 +84 val_84 2008-04-08 12 +85 val_85 2008-04-08 11 +85 val_85 2008-04-08 11 +85 val_85 2008-04-08 12 +85 val_85 2008-04-08 12 +86 val_86 2008-04-08 11 +86 val_86 2008-04-08 11 +86 val_86 2008-04-08 12 +86 val_86 2008-04-08 12 +87 val_87 2008-04-08 11 +87 val_87 2008-04-08 11 +87 val_87 2008-04-08 12 +87 val_87 2008-04-08 12 +9 val_9 2008-04-08 11 +9 val_9 2008-04-08 11 +9 val_9 2008-04-08 12 +9 val_9 2008-04-08 12 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 11 +90 val_90 2008-04-08 12 +90 val_90 2008-04-08 12 +90 val_90 2008-04-08 12 +90 val_90 2008-04-08 12 +90 val_90 2008-04-08 12 +90 val_90 2008-04-08 12 +92 val_92 2008-04-08 11 +92 val_92 2008-04-08 11 +92 val_92 2008-04-08 12 +92 val_92 2008-04-08 12 +95 val_95 2008-04-08 11 +95 val_95 2008-04-08 11 +95 val_95 2008-04-08 11 +95 val_95 2008-04-08 11 +95 val_95 2008-04-08 12 +95 val_95 2008-04-08 12 +95 val_95 2008-04-08 12 +95 val_95 2008-04-08 12 +96 val_96 2008-04-08 11 +96 val_96 2008-04-08 11 +96 val_96 2008-04-08 12 +96 val_96 2008-04-08 12 +97 val_97 2008-04-08 11 +97 val_97 2008-04-08 11 +97 val_97 2008-04-08 11 +97 val_97 2008-04-08 11 +97 val_97 2008-04-08 12 +97 val_97 2008-04-08 12 +97 val_97 2008-04-08 12 +97 val_97 2008-04-08 12 +98 val_98 2008-04-08 11 +98 val_98 2008-04-08 11 +98 val_98 2008-04-08 11 +98 val_98 2008-04-08 11 +98 val_98 2008-04-08 12 +98 val_98 2008-04-08 12 +98 val_98 2008-04-08 12 +98 val_98 2008-04-08 12 diff --git ql/src/test/results/clientpositive/input_part8.q.out ql/src/test/results/clientpositive/llap/input_part8.q.out similarity index 85% rename from ql/src/test/results/clientpositive/input_part8.q.out rename to ql/src/test/results/clientpositive/llap/input_part8.q.out index 84b1a5f161..60e6c547e1 100644 --- ql/src/test/results/clientpositive/input_part8.q.out +++ ql/src/test/results/clientpositive/llap/input_part8.q.out @@ -23,14 +23,11 @@ STAGE PLANS: TableScan alias: x filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 10 - Statistics: Num rows: 10 Data size: 4560 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE ds = '2008-04-08' LIMIT 10 diff --git ql/src/test/results/clientpositive/input_part9.q.out ql/src/test/results/clientpositive/llap/input_part9.q.out similarity index 92% rename from ql/src/test/results/clientpositive/input_part9.q.out rename to ql/src/test/results/clientpositive/llap/input_part9.q.out index 98293b1469..ec2c8930a6 100644 --- ql/src/test/results/clientpositive/input_part9.q.out +++ ql/src/test/results/clientpositive/llap/input_part9.q.out @@ -16,57 +16,14 @@ OPTIMIZED SQL: SELECT `key`, `value`, CAST('2008-04-08' AS STRING) AS `ds`, `hr` FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' AND `key` IS NOT NULL STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - filterExpr: ((ds = '2008-04-08') and key is not null) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=11 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -113,9 +70,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart -#### A masked pattern was here #### Partition - base file name: hr=12 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -162,15 +117,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [x] - /srcpart/ds=2008-04-08/hr=12 [x] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: x + filterExpr: ((ds = '2008-04-08') and key is not null) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Select Operator + expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: SELECT x.* FROM SRCPART x WHERE key IS NOT NULL AND ds = '2008-04-08' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/input_testsequencefile.q.out ql/src/test/results/clientpositive/llap/input_testsequencefile.q.out similarity index 67% rename from ql/src/test/results/clientpositive/input_testsequencefile.q.out rename to ql/src/test/results/clientpositive/llap/input_testsequencefile.q.out index 6ec8c09559..e882292886 100644 --- ql/src/test/results/clientpositive/input_testsequencefile.q.out +++ ql/src/test/results/clientpositive/llap/input_testsequencefile.q.out @@ -20,70 +20,70 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@dest4_sequencefile STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest4_sequencefile - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest4_sequencefile + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -95,7 +95,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest4_sequencefile - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -103,36 +103,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest4_sequencefile - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest4_sequencefile - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest4_sequencefile - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: FROM src INSERT OVERWRITE TABLE dest4_sequencefile SELECT src.key, src.value PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/input_testxpath.q.out ql/src/test/results/clientpositive/llap/input_testxpath.q.out new file mode 100644 index 0000000000..f5f76d9a6d --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_testxpath.q.out @@ -0,0 +1,137 @@ +PREHOOK: query: CREATE TABLE dest1_n6(key INT, value STRING, mapvalue STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n6 +POSTHOOK: query: CREATE TABLE dest1_n6(key INT, value STRING, mapvalue STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n6 +PREHOOK: query: EXPLAIN +FROM src_thrift +INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +PREHOOK: Output: default@dest1_n6 +POSTHOOK: query: EXPLAIN +FROM src_thrift +INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +POSTHOOK: Output: default@dest1_n6 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: lint[1] (type: int), lintstring[0].mystring (type: string), mstringstring['key_2'] (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n6 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, mapvalue + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(mapvalue, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 4904 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 4904 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n6 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value, mapvalue + Column Types: int, string, string + Table: default.dest1_n6 + +PREHOOK: query: FROM src_thrift +INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +PREHOOK: Output: default@dest1_n6 +POSTHOOK: query: FROM src_thrift +INSERT OVERWRITE TABLE dest1_n6 SELECT src_thrift.lint[1], src_thrift.lintstring[0].mystring, src_thrift.mstringstring['key_2'] +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +POSTHOOK: Output: default@dest1_n6 +POSTHOOK: Lineage: dest1_n6.key EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lint, type:array, comment:from deserializer), ] +POSTHOOK: Lineage: dest1_n6.mapvalue EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:mstringstring, type:map, comment:from deserializer), ] +POSTHOOK: Lineage: dest1_n6.value EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lintstring, type:array>, comment:from deserializer), ] +PREHOOK: query: SELECT dest1_n6.* FROM dest1_n6 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n6 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n6.* FROM dest1_n6 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n6 +#### A masked pattern was here #### +0 0 NULL +2 1 NULL +4 8 value_2 +6 27 NULL +8 64 NULL +10 125 NULL +12 216 NULL +14 343 NULL +16 512 NULL +18 729 NULL +NULL NULL NULL diff --git ql/src/test/results/clientpositive/llap/input_testxpath2.q.out ql/src/test/results/clientpositive/llap/input_testxpath2.q.out new file mode 100644 index 0000000000..017206a37a --- /dev/null +++ ql/src/test/results/clientpositive/llap/input_testxpath2.q.out @@ -0,0 +1,140 @@ +PREHOOK: query: CREATE TABLE dest1_n32(lint_size INT, lintstring_size INT, mstringstring_size INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n32 +POSTHOOK: query: CREATE TABLE dest1_n32(lint_size INT, lintstring_size INT, mstringstring_size INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n32 +PREHOOK: query: EXPLAIN +FROM src_thrift +INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +PREHOOK: Output: default@dest1_n32 +POSTHOOK: query: EXPLAIN +FROM src_thrift +INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +POSTHOOK: Output: default@dest1_n32 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_thrift + filterExpr: (lint is not null and mstringstring is not null) (type: boolean) + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (lint is not null and mstringstring is not null) (type: boolean) + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: size(lint) (type: int), size(lintstring) (type: int), size(mstringstring) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n32 + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: lint_size, lintstring_size, mstringstring_size + Statistics: Num rows: 11 Data size: 39600 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(lint_size, 'hll'), compute_stats(lintstring_size, 'hll'), compute_stats(mstringstring_size, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 4872 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 4872 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 4920 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n32 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: lint_size, lintstring_size, mstringstring_size + Column Types: int, int, int + Table: default.dest1_n32 + +PREHOOK: query: FROM src_thrift +INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +PREHOOK: Output: default@dest1_n32 +POSTHOOK: query: FROM src_thrift +INSERT OVERWRITE TABLE dest1_n32 SELECT size(src_thrift.lint), size(src_thrift.lintstring), size(src_thrift.mstringstring) where src_thrift.lint IS NOT NULL AND NOT (src_thrift.mstringstring IS NULL) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +POSTHOOK: Output: default@dest1_n32 +POSTHOOK: Lineage: dest1_n32.lint_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lint, type:array, comment:from deserializer), ] +POSTHOOK: Lineage: dest1_n32.lintstring_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:lintstring, type:array>, comment:from deserializer), ] +POSTHOOK: Lineage: dest1_n32.mstringstring_size EXPRESSION [(src_thrift)src_thrift.FieldSchema(name:mstringstring, type:map, comment:from deserializer), ] +PREHOOK: query: SELECT dest1_n32.* FROM dest1_n32 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n32 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n32.* FROM dest1_n32 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n32 +#### A masked pattern was here #### +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 +3 1 1 diff --git ql/src/test/results/clientpositive/input_testxpath3.q.out ql/src/test/results/clientpositive/llap/input_testxpath3.q.out similarity index 50% rename from ql/src/test/results/clientpositive/input_testxpath3.q.out rename to ql/src/test/results/clientpositive/llap/input_testxpath3.q.out index 899017a51a..62275b32f4 100644 --- ql/src/test/results/clientpositive/input_testxpath3.q.out +++ ql/src/test/results/clientpositive/llap/input_testxpath3.q.out @@ -11,33 +11,19 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src_thrift #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src_thrift + Select Operator + expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: FROM src_thrift SELECT src_thrift.mstringstring['key_9'], src_thrift.lintstring.myint diff --git ql/src/test/results/clientpositive/input_testxpath4.q.out ql/src/test/results/clientpositive/llap/input_testxpath4.q.out similarity index 51% rename from ql/src/test/results/clientpositive/input_testxpath4.q.out rename to ql/src/test/results/clientpositive/llap/input_testxpath4.q.out index a60534df41..dec59379b4 100644 --- ql/src/test/results/clientpositive/input_testxpath4.q.out +++ ql/src/test/results/clientpositive/llap/input_testxpath4.q.out @@ -17,36 +17,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src_thrift #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src_thrift + Filter Operator + predicate: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) + Select Operator + expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: FROM src_thrift SELECT src_thrift.mstringstring['key_9'], lintstring.myint @@ -93,37 +78,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src_thrift #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_thrift - filterExpr: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 11 Data size: 38280 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src_thrift + filterExpr: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) + Filter Operator + predicate: (mstringstring['key_9'] is not null and lintstring.myint is not null and lintstring is not null) (type: boolean) + Select Operator + expressions: mstringstring['key_9'] (type: string), lintstring.myint (type: array) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: FROM src_thrift SELECT src_thrift.mstringstring['key_9'], lintstring.myint diff --git ql/src/test/results/clientpositive/insert0.q.out ql/src/test/results/clientpositive/llap/insert0.q.out similarity index 83% rename from ql/src/test/results/clientpositive/insert0.q.out rename to ql/src/test/results/clientpositive/llap/insert0.q.out index 4c1fa2f14f..71579d29d7 100644 --- ql/src/test/results/clientpositive/insert0.q.out +++ ql/src/test/results/clientpositive/llap/insert0.q.out @@ -84,12 +84,12 @@ POSTHOOK: Input: default@insert_into1_n1 103 val_103 104 val_104 104 val_104 -PREHOOK: query: create table ctas_table as SELECT key, count(value) as foo from src GROUP BY key LIMIT 10 +PREHOOK: query: CREATE TABLE ctas_table AS SELECT key, count(value) AS foo FROM src GROUP BY key ORDER BY key LIMIT 10 PREHOOK: type: CREATETABLE_AS_SELECT PREHOOK: Input: default@src PREHOOK: Output: database:default PREHOOK: Output: default@ctas_table -POSTHOOK: query: create table ctas_table as SELECT key, count(value) as foo from src GROUP BY key LIMIT 10 +POSTHOOK: query: CREATE TABLE ctas_table AS SELECT key, count(value) AS foo FROM src GROUP BY key ORDER BY key LIMIT 10 POSTHOOK: type: CREATETABLE_AS_SELECT POSTHOOK: Input: default@src POSTHOOK: Output: database:default @@ -133,57 +133,47 @@ POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@ctas_part PREHOOK: query: insert overwrite table ctas_part partition (modkey) -select key, value, ceil(key / 100) from src where key is not null limit 10 +select key, value, ceil(key / 100) from src where key is not null order by key limit 10 PREHOOK: type: QUERY PREHOOK: Input: default@src PREHOOK: Output: default@ctas_part POSTHOOK: query: insert overwrite table ctas_part partition (modkey) -select key, value, ceil(key / 100) from src where key is not null limit 10 +select key, value, ceil(key / 100) from src where key is not null order by key limit 10 POSTHOOK: type: QUERY POSTHOOK: Input: default@src +POSTHOOK: Output: default@ctas_part@modkey=0 POSTHOOK: Output: default@ctas_part@modkey=1 POSTHOOK: Output: default@ctas_part@modkey=2 -POSTHOOK: Output: default@ctas_part@modkey=3 -POSTHOOK: Output: default@ctas_part@modkey=4 -POSTHOOK: Output: default@ctas_part@modkey=5 +POSTHOOK: Lineage: ctas_part PARTITION(modkey=0).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: ctas_part PARTITION(modkey=0).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: ctas_part PARTITION(modkey=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: ctas_part PARTITION(modkey=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] POSTHOOK: Lineage: ctas_part PARTITION(modkey=2).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: ctas_part PARTITION(modkey=2).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=3).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=3).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=4).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=4).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=5).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: ctas_part PARTITION(modkey=5).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: select * from ctas_part order by key PREHOOK: type: QUERY PREHOOK: Input: default@ctas_part +PREHOOK: Input: default@ctas_part@modkey=0 PREHOOK: Input: default@ctas_part@modkey=1 PREHOOK: Input: default@ctas_part@modkey=2 -PREHOOK: Input: default@ctas_part@modkey=3 -PREHOOK: Input: default@ctas_part@modkey=4 -PREHOOK: Input: default@ctas_part@modkey=5 #### A masked pattern was here #### POSTHOOK: query: select * from ctas_part order by key POSTHOOK: type: QUERY POSTHOOK: Input: default@ctas_part +POSTHOOK: Input: default@ctas_part@modkey=0 POSTHOOK: Input: default@ctas_part@modkey=1 POSTHOOK: Input: default@ctas_part@modkey=2 -POSTHOOK: Input: default@ctas_part@modkey=3 -POSTHOOK: Input: default@ctas_part@modkey=4 -POSTHOOK: Input: default@ctas_part@modkey=5 #### A masked pattern was here #### -27 val_27 1 -86 val_86 1 -98 val_98 1 -165 val_165 2 -238 val_238 3 -255 val_255 3 -278 val_278 3 -311 val_311 4 -409 val_409 5 -484 val_484 5 +0 val_0 0 +0 val_0 0 +0 val_0 0 +10 val_10 1 +100 val_100 1 +100 val_100 1 +103 val_103 2 +103 val_103 2 +104 val_104 2 +104 val_104 2 PREHOOK: query: DROP TABLE insert_into1_n1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1_n1 diff --git ql/src/test/results/clientpositive/llap/insert2_overwrite_partitions.q.out ql/src/test/results/clientpositive/llap/insert2_overwrite_partitions.q.out new file mode 100644 index 0000000000..f94d6f6efb --- /dev/null +++ ql/src/test/results/clientpositive/llap/insert2_overwrite_partitions.q.out @@ -0,0 +1,397 @@ +PREHOOK: query: CREATE DATABASE db1 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:db1 +POSTHOOK: query: CREATE DATABASE db1 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:db1 +PREHOOK: query: CREATE DATABASE db2 +PREHOOK: type: CREATEDATABASE +PREHOOK: Output: database:db2 +POSTHOOK: query: CREATE DATABASE db2 +POSTHOOK: type: CREATEDATABASE +POSTHOOK: Output: database:db2 +PREHOOK: query: CREATE TABLE db1.sourceTable (one string,two string) PARTITIONED BY (ds string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:db1 +PREHOOK: Output: db1@sourceTable +POSTHOOK: query: CREATE TABLE db1.sourceTable (one string,two string) PARTITIONED BY (ds string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:db1 +POSTHOOK: Output: db1@sourceTable +PREHOOK: query: load data local inpath '../../data/files/kv1.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: db1@sourcetable +POSTHOOK: query: load data local inpath '../../data/files/kv1.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: db1@sourcetable +POSTHOOK: Output: db1@sourcetable@ds=2011-11-11 +PREHOOK: query: load data local inpath '../../data/files/kv3.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: db1@sourcetable@ds=2011-11-11 +POSTHOOK: query: load data local inpath '../../data/files/kv3.txt' INTO TABLE db1.sourceTable partition(ds='2011-11-11') +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: db1@sourcetable@ds=2011-11-11 +PREHOOK: query: CREATE TABLE db2.destinTable (one string,two string) PARTITIONED BY (ds string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:db2 +PREHOOK: Output: db2@destinTable +POSTHOOK: query: CREATE TABLE db2.destinTable (one string,two string) PARTITIONED BY (ds string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:db2 +POSTHOOK: Output: db2@destinTable +PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: db1@sourcetable +PREHOOK: Input: db1@sourcetable@ds=2011-11-11 +PREHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: db1@sourcetable +POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 +POSTHOOK: Output: db2@destintable@ds=2011-11-11 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: sourcetable + filterExpr: (ds = '2011-11-11') (type: boolean) + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Top N Key Operator + sort order: -- + keys: one (type: string), two (type: string) + null sort order: zz + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + top n: 5 + Select Operator + expressions: one (type: string), two (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -- + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: db2.destintable + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2011-11-11' (type: string) + outputColumnNames: one, two, ds + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: compute_stats(one, 'hll'), compute_stats(two, 'hll') + keys: ds (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + value expressions: _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2011-11-11 + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: db2.destintable + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: one, two + Column Types: string, string + Table: db2.destintable + +PREHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: db1@sourcetable +PREHOOK: Input: db1@sourcetable@ds=2011-11-11 +PREHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: db1@sourcetable +POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 +POSTHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).one SIMPLE [(sourcetable)sourcetable.FieldSchema(name:one, type:string, comment:null), ] +POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).two SIMPLE [(sourcetable)sourcetable.FieldSchema(name:two, type:string, comment:null), ] +PREHOOK: query: select one,two from db2.destinTable order by one desc, two desc +PREHOOK: type: QUERY +PREHOOK: Input: db2@destintable +PREHOOK: Input: db2@destintable@ds=2011-11-11 +#### A masked pattern was here #### +POSTHOOK: query: select one,two from db2.destinTable order by one desc, two desc +POSTHOOK: type: QUERY +POSTHOOK: Input: db2@destintable +POSTHOOK: Input: db2@destintable@ds=2011-11-11 +#### A masked pattern was here #### +98 val_98 +98 val_98 +98 val_98 +97 val_97 +97 val_97 +PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: db1@sourcetable +PREHOOK: Input: db1@sourcetable@ds=2011-11-11 +PREHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: db1@sourcetable +POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 +POSTHOOK: Output: db2@destintable@ds=2011-11-11 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: sourcetable + filterExpr: (ds = '2011-11-11') (type: boolean) + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Top N Key Operator + sort order: -- + keys: one (type: string), two (type: string) + null sort order: zz + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + top n: 5 + Select Operator + expressions: one (type: string), two (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -- + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 124 Data size: 40480 Basic stats: PARTIAL Column stats: NONE + Limit + Number of rows: 5 + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: db2.destintable + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2011-11-11' (type: string) + outputColumnNames: one, two, ds + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: compute_stats(one, 'hll'), compute_stats(two, 'hll') + keys: ds (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 5 Data size: 1630 Basic stats: PARTIAL Column stats: NONE + value expressions: _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 652 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2011-11-11 + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: db2.destintable + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: one, two + Column Types: string, string + Table: db2.destintable + +PREHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +PREHOOK: type: QUERY +PREHOOK: Input: db1@sourcetable +PREHOOK: Input: db1@sourcetable@ds=2011-11-11 +PREHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: query: INSERT OVERWRITE TABLE db2.destinTable PARTITION (ds='2011-11-11') +SELECT one,two FROM db1.sourceTable WHERE ds='2011-11-11' order by one desc, two desc limit 5 +POSTHOOK: type: QUERY +POSTHOOK: Input: db1@sourcetable +POSTHOOK: Input: db1@sourcetable@ds=2011-11-11 +POSTHOOK: Output: db2@destintable@ds=2011-11-11 +POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).one SIMPLE [(sourcetable)sourcetable.FieldSchema(name:one, type:string, comment:null), ] +POSTHOOK: Lineage: destintable PARTITION(ds=2011-11-11).two SIMPLE [(sourcetable)sourcetable.FieldSchema(name:two, type:string, comment:null), ] +PREHOOK: query: select one,two from db2.destinTable order by one desc, two desc +PREHOOK: type: QUERY +PREHOOK: Input: db2@destintable +PREHOOK: Input: db2@destintable@ds=2011-11-11 +#### A masked pattern was here #### +POSTHOOK: query: select one,two from db2.destinTable order by one desc, two desc +POSTHOOK: type: QUERY +POSTHOOK: Input: db2@destintable +POSTHOOK: Input: db2@destintable@ds=2011-11-11 +#### A masked pattern was here #### +98 val_98 +98 val_98 +98 val_98 +97 val_97 +97 val_97 +PREHOOK: query: drop table db2.destinTable +PREHOOK: type: DROPTABLE +PREHOOK: Input: db2@destintable +PREHOOK: Output: db2@destintable +POSTHOOK: query: drop table db2.destinTable +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: db2@destintable +POSTHOOK: Output: db2@destintable +PREHOOK: query: drop table db1.sourceTable +PREHOOK: type: DROPTABLE +PREHOOK: Input: db1@sourcetable +PREHOOK: Output: db1@sourcetable +POSTHOOK: query: drop table db1.sourceTable +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: db1@sourcetable +POSTHOOK: Output: db1@sourcetable +PREHOOK: query: DROP DATABASE db1 +PREHOOK: type: DROPDATABASE +PREHOOK: Input: database:db1 +PREHOOK: Output: database:db1 +POSTHOOK: query: DROP DATABASE db1 +POSTHOOK: type: DROPDATABASE +POSTHOOK: Input: database:db1 +POSTHOOK: Output: database:db1 +PREHOOK: query: DROP DATABASE db2 +PREHOOK: type: DROPDATABASE +PREHOOK: Input: database:db2 +PREHOOK: Output: database:db2 +POSTHOOK: query: DROP DATABASE db2 +POSTHOOK: type: DROPDATABASE +POSTHOOK: Input: database:db2 +POSTHOOK: Output: database:db2 diff --git ql/src/test/results/clientpositive/insert_nonacid_from_acid.q.out ql/src/test/results/clientpositive/llap/insert_nonacid_from_acid.q.out similarity index 97% rename from ql/src/test/results/clientpositive/insert_nonacid_from_acid.q.out rename to ql/src/test/results/clientpositive/llap/insert_nonacid_from_acid.q.out index daf466e5ba..b6108c6d9c 100644 --- ql/src/test/results/clientpositive/insert_nonacid_from_acid.q.out +++ ql/src/test/results/clientpositive/llap/insert_nonacid_from_acid.q.out @@ -25,11 +25,11 @@ POSTHOOK: query: select * from sample_06 where gpa = 3.00 POSTHOOK: type: QUERY POSTHOOK: Input: default@sample_06 #### A masked pattern was here #### -aaa 35 3.00 -bbb 32 3.00 -ccc 32 3.00 -ddd 35 3.00 -eee 32 3.00 +aaa 35 3 +bbb 32 3 +ccc 32 3 +ddd 35 3 +eee 32 3 PREHOOK: query: create table tab1_n2 (name varchar(50), age int, gpa decimal(3, 2)) PREHOOK: type: CREATETABLE PREHOOK: Output: database:default diff --git ql/src/test/results/clientpositive/llap/insertoverwrite_bucket.q.out ql/src/test/results/clientpositive/llap/insertoverwrite_bucket.q.out new file mode 100644 index 0000000000..54a7ac0d49 --- /dev/null +++ ql/src/test/results/clientpositive/llap/insertoverwrite_bucket.q.out @@ -0,0 +1,379 @@ +PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketinput( +data string +) +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@bucketinput +POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketinput( +data string +) +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@bucketinput +PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput1( +data string +)CLUSTERED BY(data) +INTO 2 BUCKETS +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@bucketoutput1 +POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput1( +data string +)CLUSTERED BY(data) +INTO 2 BUCKETS +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@bucketoutput1 +PREHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput2( +data string +)CLUSTERED BY(data) +INTO 2 BUCKETS +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@bucketoutput2 +POSTHOOK: query: CREATE TABLE IF NOT EXISTS bucketoutput2( +data string +)CLUSTERED BY(data) +INTO 2 BUCKETS +ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@bucketoutput2 +PREHOOK: query: insert into table bucketinput values ("firstinsert1") +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@bucketinput +POSTHOOK: query: insert into table bucketinput values ("firstinsert1") +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@bucketinput +POSTHOOK: Lineage: bucketinput.data SCRIPT [] +PREHOOK: query: insert into table bucketinput values ("firstinsert2") +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@bucketinput +POSTHOOK: query: insert into table bucketinput values ("firstinsert2") +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@bucketinput +POSTHOOK: Lineage: bucketinput.data SCRIPT [] +PREHOOK: query: insert into table bucketinput values ("firstinsert3") +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@bucketinput +POSTHOOK: query: insert into table bucketinput values ("firstinsert3") +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@bucketinput +POSTHOOK: Lineage: bucketinput.data SCRIPT [] +PREHOOK: query: insert overwrite table bucketoutput1 select * from bucketinput where data like 'first%' +PREHOOK: type: QUERY +PREHOOK: Input: default@bucketinput +PREHOOK: Output: default@bucketoutput1 +POSTHOOK: query: insert overwrite table bucketoutput1 select * from bucketinput where data like 'first%' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@bucketinput +POSTHOOK: Output: default@bucketoutput1 +POSTHOOK: Lineage: bucketoutput1.data SIMPLE [(bucketinput)bucketinput.FieldSchema(name:data, type:string, comment:null), ] +PREHOOK: query: CREATE TABLE temp1 +( + change string, + num string +) +CLUSTERED BY (num) SORTED BY (num) INTO 4 BUCKETS +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@temp1 +POSTHOOK: query: CREATE TABLE temp1 +( + change string, + num string +) +CLUSTERED BY (num) SORTED BY (num) INTO 4 BUCKETS +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@temp1 +PREHOOK: query: explain insert overwrite table temp1 select data, data from bucketinput +PREHOOK: type: QUERY +PREHOOK: Input: default@bucketinput +PREHOOK: Output: default@temp1 +POSTHOOK: query: explain insert overwrite table temp1 select data, data from bucketinput +POSTHOOK: type: QUERY +POSTHOOK: Input: default@bucketinput +POSTHOOK: Output: default@temp1 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: bucketinput + Statistics: Num rows: 3 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: data (type: string) + outputColumnNames: _col1 + Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.temp1 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: change, num + Statistics: Num rows: 3 Data size: 576 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(change, 'hll'), compute_stats(num, 'hll') + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.temp1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: change, num + Column Types: string, string + Table: default.temp1 + +PREHOOK: query: CREATE TABLE temp2 +( + create_ts STRING , + change STRING, + num STRING +) +CLUSTERED BY (create_ts) SORTED BY (num) INTO 4 BUCKETS +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@temp2 +POSTHOOK: query: CREATE TABLE temp2 +( + create_ts STRING , + change STRING, + num STRING +) +CLUSTERED BY (create_ts) SORTED BY (num) INTO 4 BUCKETS +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@temp2 +PREHOOK: query: explain +INSERT OVERWRITE TABLE temp2 +SELECT change, change,num +FROM temp1 +PREHOOK: type: QUERY +PREHOOK: Input: default@temp1 +PREHOOK: Output: default@temp2 +POSTHOOK: query: explain +INSERT OVERWRITE TABLE temp2 +SELECT change, change,num +FROM temp1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@temp1 +POSTHOOK: Output: default@temp2 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: temp1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: change (type: string), num (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.temp2 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: create_ts, change, num + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: compute_stats(create_ts, 'hll'), compute_stats(change, 'hll'), compute_stats(num, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1688 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.temp2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: create_ts, change, num + Column Types: string, string, string + Table: default.temp2 + +PREHOOK: query: select * from bucketoutput1 a join bucketoutput2 b on (a.data=b.data) +PREHOOK: type: QUERY +PREHOOK: Input: default@bucketoutput1 +PREHOOK: Input: default@bucketoutput2 +#### A masked pattern was here #### +POSTHOOK: query: select * from bucketoutput1 a join bucketoutput2 b on (a.data=b.data) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@bucketoutput1 +POSTHOOK: Input: default@bucketoutput2 +#### A masked pattern was here #### +PREHOOK: query: drop table temp1 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@temp1 +PREHOOK: Output: default@temp1 +POSTHOOK: query: drop table temp1 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@temp1 +POSTHOOK: Output: default@temp1 +PREHOOK: query: drop table temp2 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@temp2 +PREHOOK: Output: default@temp2 +POSTHOOK: query: drop table temp2 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@temp2 +POSTHOOK: Output: default@temp2 +PREHOOK: query: drop table buckettestinput +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table buckettestinput +POSTHOOK: type: DROPTABLE +PREHOOK: query: drop table buckettestoutput1 +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table buckettestoutput1 +POSTHOOK: type: DROPTABLE +PREHOOK: query: drop table buckettestoutput2 +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table buckettestoutput2 +POSTHOOK: type: DROPTABLE diff --git ql/src/test/results/clientpositive/intersect_all_rj.q.out ql/src/test/results/clientpositive/llap/intersect_all_rj.q.out similarity index 98% rename from ql/src/test/results/clientpositive/intersect_all_rj.q.out rename to ql/src/test/results/clientpositive/llap/intersect_all_rj.q.out index a807f11301..5647eb4953 100644 --- ql/src/test/results/clientpositive/intersect_all_rj.q.out +++ ql/src/test/results/clientpositive/llap/intersect_all_rj.q.out @@ -109,7 +109,7 @@ POSTHOOK: Lineage: table_10.int_col_18 SCRIPT [] POSTHOOK: Lineage: table_10.timestamp_col_15 SCRIPT [] POSTHOOK: Lineage: table_10.timestamp_col_30 SCRIPT [] POSTHOOK: Lineage: table_10.timestamp_col_5 SCRIPT [] -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join MERGEJOIN[47][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: explain cbo SELECT DISTINCT COALESCE(a4.timestamp_col_15, IF(a4.boolean_col_16, a4.timestamp_col_30, a4.timestamp_col_5)) AS timestamp_col @@ -194,7 +194,7 @@ HiveProject($f0=[$1]) HiveProject($f0=[CASE(IS NOT NULL(least(CASE(IS NOT NULL($0), $0, 2010-03-29 00:00:00:TIMESTAMP(9)), CASE(IS NOT NULL($1), $1, 2014-08-16 00:00:00:TIMESTAMP(9)))), least(CASE(IS NOT NULL($0), $0, 2010-03-29 00:00:00:TIMESTAMP(9)), CASE(IS NOT NULL($1), $1, 2014-08-16 00:00:00:TIMESTAMP(9))), greatest(CASE(IS NOT NULL($0), $0, 2013-07-01 00:00:00:TIMESTAMP(9)), CASE(IS NOT NULL($1), $1, 2028-06-18 00:00:00:TIMESTAMP(9))))]) HiveTableScan(table=[[default, table_16]], table:alias=[a1]) -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join MERGEJOIN[47][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: SELECT DISTINCT COALESCE(a4.timestamp_col_15, IF(a4.boolean_col_16, a4.timestamp_col_30, a4.timestamp_col_5)) AS timestamp_col FROM table_7 a3 @@ -251,5 +251,5 @@ POSTHOOK: Input: default@table_10 POSTHOOK: Input: default@table_16 POSTHOOK: Input: default@table_7 #### A masked pattern was here #### -2018-01-10 15:03:55 2018-02-10 07:12:55 +2018-01-10 15:03:55 diff --git ql/src/test/results/clientpositive/interval_3.q.out ql/src/test/results/clientpositive/llap/interval_3.q.out similarity index 58% rename from ql/src/test/results/clientpositive/interval_3.q.out rename to ql/src/test/results/clientpositive/llap/interval_3.q.out index e2c43118df..dc8b3930cf 100644 --- ql/src/test/results/clientpositive/interval_3.q.out +++ ql/src/test/results/clientpositive/llap/interval_3.q.out @@ -15,9 +15,9 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@lineitem #### A masked pattern was here #### 6 1992-04-27 1992-05-02 -37 1992-07-10 1992-08-02 -37 1992-07-02 1992-07-28 37 1992-07-21 1992-08-15 +37 1992-07-02 1992-07-28 +37 1992-07-10 1992-08-02 PREHOOK: query: select l_orderkey, l_shipdate, l_receiptdate from lineitem @@ -35,9 +35,9 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@lineitem #### A masked pattern was here #### 6 1992-04-27 1992-05-02 -37 1992-07-10 1992-08-02 -37 1992-07-02 1992-07-28 37 1992-07-21 1992-08-15 +37 1992-07-02 1992-07-28 +37 1992-07-10 1992-08-02 PREHOOK: query: select l_orderkey, l_shipdate, l_receiptdate from lineitem @@ -108,7 +108,7 @@ POSTHOOK: query: create table date_dim_d1( POSTHOOK: type: CREATETABLE POSTHOOK: Output: database:default POSTHOOK: Output: default@date_dim_d1 -Warning: Shuffle Join JOIN[8][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product +Warning: Shuffle Join MERGEJOIN[16][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: EXPLAIN SELECT d1.d_week_seq FROM @@ -149,63 +149,75 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: d1 - filterExpr: CAST( d_date AS DATE) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: CAST( d_date AS DATE) is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: d_week_seq (type: int), ((((((CAST( d_date AS DATE) + INTERVAL'1-0') + INTERVAL'0-2') + INTERVAL'5 00:00:00.000000000') + INTERVAL'0 04:00:00.000000000') + INTERVAL'0 00:10:00.000000000') + INTERVAL'0 00:00:09.000000000') (type: timestamp), (CAST( d_date AS DATE) + INTERVAL'1-2') (type: date) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: d1 + filterExpr: CAST( d_date AS DATE) is not null (type: boolean) Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: timestamp), _col2 (type: date) - TableScan - alias: d3 - filterExpr: (CAST( CAST( d_date AS DATE) AS TIMESTAMP) is not null and CAST( d_date AS DATE) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (CAST( CAST( d_date AS DATE) AS TIMESTAMP) is not null and CAST( d_date AS DATE) is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: CAST( CAST( d_date AS DATE) AS TIMESTAMP) (type: timestamp), CAST( d_date AS DATE) (type: date) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: + Filter Operator + predicate: CAST( d_date AS DATE) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: d_week_seq (type: int), ((((((CAST( d_date AS DATE) + INTERVAL'1-0') + INTERVAL'0-2') + INTERVAL'5 00:00:00.000000000') + INTERVAL'0 04:00:00.000000000') + INTERVAL'0 00:10:00.000000000') + INTERVAL'0 00:00:09.000000000') (type: timestamp), (CAST( d_date AS DATE) + INTERVAL'1-2') (type: date) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: timestamp), _col2 (type: date) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: d3 + filterExpr: (CAST( CAST( d_date AS DATE) AS TIMESTAMP) is not null and CAST( d_date AS DATE) is not null) (type: boolean) Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: timestamp), _col1 (type: date) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((_col4 < _col2) and (_col3 > _col1)) (type: boolean) - Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Filter Operator + predicate: (CAST( CAST( d_date AS DATE) AS TIMESTAMP) is not null and CAST( d_date AS DATE) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: CAST( CAST( d_date AS DATE) AS TIMESTAMP) (type: timestamp), CAST( d_date AS DATE) (type: date) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: timestamp), _col1 (type: date) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + residual filter predicates: {(_col3 > _col1)} {(_col4 < _col2)} Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 373 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/interval_alt.q.out ql/src/test/results/clientpositive/llap/interval_alt.q.out similarity index 79% rename from ql/src/test/results/clientpositive/interval_alt.q.out rename to ql/src/test/results/clientpositive/llap/interval_alt.q.out index 7b5387ba12..550091a74d 100644 --- ql/src/test/results/clientpositive/interval_alt.q.out +++ ql/src/test/results/clientpositive/llap/interval_alt.q.out @@ -132,34 +132,19 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@t_n18 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: t_n18 - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: (DATE'2012-01-01' + IntervalDayLiteralProcessor(((- dt) * dt))) (type: timestamp), (DATE'2012-01-01' - IntervalDayLiteralProcessor(((- dt) * dt))) (type: timestamp), TIMESTAMP'2012-01-04 00:00:00' (type: timestamp), (DATE'2012-01-01' + IntervalYearMonthLiteralProcessor(concat(CAST( dt AS STRING), '-1'))) (type: date) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 352 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 352 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: t_n18 + Select Operator + expressions: (DATE'2012-01-01' + IntervalDayLiteralProcessor(((- dt) * dt))) (type: timestamp), (DATE'2012-01-01' - IntervalDayLiteralProcessor(((- dt) * dt))) (type: timestamp), TIMESTAMP'2012-01-04 00:00:00' (type: timestamp), (DATE'2012-01-01' + IntervalYearMonthLiteralProcessor(concat(CAST( dt AS STRING), '-1'))) (type: date) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select date '2012-01-01' + interval (-dt*dt) day, diff --git ql/src/test/results/clientpositive/interval_arithmetic.q.out ql/src/test/results/clientpositive/llap/interval_arithmetic.q.out similarity index 65% rename from ql/src/test/results/clientpositive/interval_arithmetic.q.out rename to ql/src/test/results/clientpositive/llap/interval_arithmetic.q.out index 3b5db9b1ed..02a3fc5ca0 100644 --- ql/src/test/results/clientpositive/interval_arithmetic.q.out +++ ql/src/test/results/clientpositive/llap/interval_arithmetic.q.out @@ -47,37 +47,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 513744 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: dateval (type: date), (dateval - INTERVAL'2-2') (type: date), (dateval - INTERVAL'-2-2') (type: date), (dateval + INTERVAL'2-2') (type: date), (dateval + INTERVAL'-2-2') (type: date), (INTERVAL'-2-2' + dateval) (type: date), (INTERVAL'2-2' + dateval) (type: date) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 12288 Data size: 4642512 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 784 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: dateval (type: date), (dateval - INTERVAL'2-2') (type: date), (dateval - INTERVAL'-2-2') (type: date), (dateval + INTERVAL'2-2') (type: date), (dateval + INTERVAL'-2-2') (type: date), (INTERVAL'-2-2' + dateval) (type: date), (INTERVAL'2-2' + dateval) (type: date) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select dateval, @@ -130,37 +114,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 513744 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: dateval (type: date), (dateval - DATE'1999-06-07') (type: interval_day_time), (DATE'1999-06-07' - dateval) (type: interval_day_time), (dateval - dateval) (type: interval_day_time) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12288 Data size: 956112 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: dateval (type: date), (dateval - DATE'1999-06-07') (type: interval_day_time), (DATE'1999-06-07' - dateval) (type: interval_day_time), (dateval - dateval) (type: interval_day_time) + outputColumnNames: _col0, _col1, _col2, _col3 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select dateval, @@ -213,37 +181,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 366960 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: tsval (type: timestamp), (tsval - INTERVAL'2-2') (type: timestamp), (tsval - INTERVAL'-2-2') (type: timestamp), (tsval + INTERVAL'2-2') (type: timestamp), (tsval + INTERVAL'-2-2') (type: timestamp), (INTERVAL'-2-2' + tsval) (type: timestamp), (INTERVAL'2-2' + tsval) (type: timestamp) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 12288 Data size: 3316080 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: tsval (type: timestamp), (tsval - INTERVAL'2-2') (type: timestamp), (tsval - INTERVAL'-2-2') (type: timestamp), (tsval + INTERVAL'2-2') (type: timestamp), (tsval + INTERVAL'-2-2') (type: timestamp), (INTERVAL'-2-2' + tsval) (type: timestamp), (INTERVAL'2-2' + tsval) (type: timestamp) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select tsval, @@ -301,14 +253,11 @@ STAGE PLANS: Processor Tree: TableScan alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 326837 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: INTERVAL'5-5' (type: interval_year_month), INTERVAL'-1-1' (type: interval_year_month) outputColumnNames: _col0, _col1 - Statistics: Num rows: 12288 Data size: 196608 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 2 - Statistics: Num rows: 2 Data size: 32 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: select @@ -358,37 +307,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 513744 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: dateval (type: date), (dateval - INTERVAL'99 11:22:33.123456789') (type: timestamp), (dateval - INTERVAL'-99 11:22:33.123456789') (type: timestamp), (dateval + INTERVAL'99 11:22:33.123456789') (type: timestamp), (dateval + INTERVAL'-99 11:22:33.123456789') (type: timestamp), (INTERVAL'-99 11:22:33.123456789' + dateval) (type: timestamp), (INTERVAL'99 11:22:33.123456789' + dateval) (type: timestamp) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 12288 Data size: 3462864 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 592 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 592 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: dateval (type: date), (dateval - INTERVAL'99 11:22:33.123456789') (type: timestamp), (dateval - INTERVAL'-99 11:22:33.123456789') (type: timestamp), (dateval + INTERVAL'99 11:22:33.123456789') (type: timestamp), (dateval + INTERVAL'-99 11:22:33.123456789') (type: timestamp), (INTERVAL'-99 11:22:33.123456789' + dateval) (type: timestamp), (INTERVAL'99 11:22:33.123456789' + dateval) (type: timestamp) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select dateval, @@ -443,37 +376,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 880704 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: dateval (type: date), tsval (type: timestamp), (dateval - tsval) (type: interval_day_time), (tsval - dateval) (type: interval_day_time), (tsval - tsval) (type: interval_day_time) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 12288 Data size: 1323072 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 264 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: dateval (type: date), tsval (type: timestamp), (dateval - tsval) (type: interval_day_time), (tsval - dateval) (type: interval_day_time), (tsval - tsval) (type: interval_day_time) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select dateval, @@ -528,37 +445,21 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@interval_arithmetic_1_n0 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 366960 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: tsval (type: timestamp), (tsval - INTERVAL'99 11:22:33.123456789') (type: timestamp), (tsval - INTERVAL'-99 11:22:33.123456789') (type: timestamp), (tsval + INTERVAL'99 11:22:33.123456789') (type: timestamp), (tsval + INTERVAL'-99 11:22:33.123456789') (type: timestamp), (INTERVAL'-99 11:22:33.123456789' + tsval) (type: timestamp), (INTERVAL'99 11:22:33.123456789' + tsval) (type: timestamp) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 12288 Data size: 3316080 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 560 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: 2 Processor Tree: - ListSink + TableScan + alias: interval_arithmetic_1_n0 + Select Operator + expressions: tsval (type: timestamp), (tsval - INTERVAL'99 11:22:33.123456789') (type: timestamp), (tsval - INTERVAL'-99 11:22:33.123456789') (type: timestamp), (tsval + INTERVAL'99 11:22:33.123456789') (type: timestamp), (tsval + INTERVAL'-99 11:22:33.123456789') (type: timestamp), (INTERVAL'-99 11:22:33.123456789' + tsval) (type: timestamp), (INTERVAL'99 11:22:33.123456789' + tsval) (type: timestamp) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Limit + Number of rows: 2 + ListSink PREHOOK: query: select tsval, @@ -616,14 +517,11 @@ STAGE PLANS: Processor Tree: TableScan alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 326837 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: INTERVAL'109 20:30:40.246913578' (type: interval_day_time), INTERVAL'89 02:14:26.000000000' (type: interval_day_time) outputColumnNames: _col0, _col1 - Statistics: Num rows: 12288 Data size: 294912 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 2 - Statistics: Num rows: 2 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: select @@ -664,14 +562,11 @@ STAGE PLANS: Processor Tree: TableScan alias: interval_arithmetic_1_n0 - Statistics: Num rows: 12288 Data size: 326837 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: TIMESTAMP'2016-11-11 03:04:00' (type: timestamp) outputColumnNames: _col0 - Statistics: Num rows: 12288 Data size: 491520 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 40 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: select date '2016-11-08' + interval '1 2:02:00' day to second + interval '2' day + interval '1' hour + interval '1' minute + interval '60' second from interval_arithmetic_1_n0 limit 1 diff --git ql/src/test/results/clientpositive/ivyDownload.q.out ql/src/test/results/clientpositive/llap/ivyDownload.q.out similarity index 89% rename from ql/src/test/results/clientpositive/ivyDownload.q.out rename to ql/src/test/results/clientpositive/llap/ivyDownload.q.out index 4529551885..c5815ce233 100644 --- ql/src/test/results/clientpositive/ivyDownload.q.out +++ ql/src/test/results/clientpositive/llap/ivyDownload.q.out @@ -38,14 +38,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 3 (type: int), 6 (type: int), 10 (type: int), 3.3000000000000003D (type: double), 6.6D (type: double), 11.0D (type: double), 10.4D (type: double) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 500 Data size: 22000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 44 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT example_add(1, 2), diff --git ql/src/test/results/clientpositive/join10.q.out ql/src/test/results/clientpositive/llap/join10.q.out similarity index 78% rename from ql/src/test/results/clientpositive/join10.q.out rename to ql/src/test/results/clientpositive/llap/join10.q.out index 1c18ce71cc..a440c1e33a 100644 --- ql/src/test/results/clientpositive/join10.q.out +++ ql/src/test/results/clientpositive/llap/join10.q.out @@ -22,63 +22,77 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join11.q.out ql/src/test/results/clientpositive/llap/join11.q.out new file mode 100644 index 0000000000..053b8c7914 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join11.q.out @@ -0,0 +1,270 @@ +PREHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +10 val_10 +11 val_11 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +17 val_17 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +19 val_19 +2 val_2 +20 val_20 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +27 val_27 +28 val_28 +30 val_30 +33 val_33 +34 val_34 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +4 val_4 +41 val_41 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +43 val_43 +44 val_44 +47 val_47 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +80 val_80 +82 val_82 +83 val_83 +83 val_83 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +87 val_87 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +95 val_95 +95 val_95 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +97 val_97 +97 val_97 +98 val_98 +98 val_98 +98 val_98 +98 val_98 diff --git ql/src/test/results/clientpositive/llap/join12.q.out ql/src/test/results/clientpositive/llap/join12.q.out new file mode 100644 index 0000000000..eb3f923714 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join12.q.out @@ -0,0 +1,405 @@ +PREHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 = src3.c5 AND src3.c5 < 80 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 = src3.c5 AND src3.c5 < 80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 80.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 = src3.c5 AND src3.c5 < 80 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 = src3.c5 AND src3.c5 < 80 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +10 val_10 +11 val_11 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +17 val_17 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +18 val_18 +19 val_19 +2 val_2 +20 val_20 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +24 val_24 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +26 val_26 +27 val_27 +28 val_28 +30 val_30 +33 val_33 +34 val_34 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +4 val_4 +41 val_41 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +43 val_43 +44 val_44 +47 val_47 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +9 val_9 diff --git ql/src/test/results/clientpositive/llap/join13.q.out ql/src/test/results/clientpositive/llap/join13.q.out new file mode 100644 index 0000000000..7f53209c6d --- /dev/null +++ ql/src/test/results/clientpositive/llap/join13.q.out @@ -0,0 +1,330 @@ +PREHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: ((UDFToDouble(key) < 200.0D) or (UDFToDouble(key) < 100.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 200.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToDouble(key) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: double) + Statistics: Num rows: 166 Data size: 1328 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 43990 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) + Statistics: Num rows: 166 Data size: 43990 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 (UDFToDouble(_col2) + UDFToDouble(_col0)) (type: double) + 1 _col0 (type: double) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT src1.c1, src2.c4 +FROM +(SELECT src.key as c1, src.value as c2 from src) src1 +JOIN +(SELECT src.key as c3, src.value as c4 from src) src2 +ON src1.c1 = src2.c3 AND src1.c1 < 100 +JOIN +(SELECT src.key as c5, src.value as c6 from src) src3 +ON src1.c1 + src2.c3 = src3.c5 AND src3.c5 < 200 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +0 val_0 +10 val_10 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +12 val_12 +15 val_15 +15 val_15 +15 val_15 +15 val_15 +17 val_17 +2 val_2 +27 val_27 +33 val_33 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +35 val_35 +37 val_37 +37 val_37 +37 val_37 +37 val_37 +4 val_4 +41 val_41 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +42 val_42 +43 val_43 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +5 val_5 +57 val_57 +58 val_58 +58 val_58 +58 val_58 +58 val_58 +64 val_64 +64 val_64 +64 val_64 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +67 val_67 +69 val_69 +69 val_69 +69 val_69 +69 val_69 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +76 val_76 +78 val_78 +80 val_80 +82 val_82 +82 val_82 +83 val_83 +83 val_83 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +86 val_86 +87 val_87 +87 val_87 +9 val_9 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +90 val_90 +95 val_95 +95 val_95 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +97 val_97 +97 val_97 +98 val_98 +98 val_98 +98 val_98 +98 val_98 diff --git ql/src/test/results/clientpositive/join14.q.out ql/src/test/results/clientpositive/llap/join14.q.out similarity index 81% rename from ql/src/test/results/clientpositive/join14.q.out rename to ql/src/test/results/clientpositive/llap/join14.q.out index cb1c00790e..dfe667f472 100644 --- ql/src/test/results/clientpositive/join14.q.out +++ ql/src/test/results/clientpositive/llap/join14.q.out @@ -26,86 +26,118 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Output: default@dest1_n164 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) > 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) > 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: srcpart - filterExpr: ((UDFToDouble(key) > 100.0D) and (ds = '2008-04-08')) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) > 100.0D) (type: boolean) - Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 174 Data size: 30972 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n164 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: c1, c2 - Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) > 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) > 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 14442 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (UDFToDouble(key) > 100.0D) (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) > 100.0D) (type: boolean) + Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 333 Data size: 59274 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 174 Data size: 30972 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n164 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: c1, c2 + Statistics: Num rows: 174 Data size: 16530 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -117,7 +149,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n164 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -125,30 +157,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n164 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM src JOIN srcpart ON src.key = srcpart.key AND srcpart.ds = '2008-04-08' and src.key > 100 INSERT OVERWRITE TABLE dest1_n164 SELECT src.key, srcpart.value PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/join15.q.out ql/src/test/results/clientpositive/llap/join15.q.out similarity index 85% rename from ql/src/test/results/clientpositive/join15.q.out rename to ql/src/test/results/clientpositive/llap/join15.q.out index bbde586a2a..4735a49d97 100644 --- ql/src/test/results/clientpositive/join15.q.out +++ ql/src/test/results/clientpositive/llap/join15.q.out @@ -10,87 +10,91 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - null sort order: zzzz - sort order: ++++ - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join16.q.out ql/src/test/results/clientpositive/llap/join16.q.out new file mode 100644 index 0000000000..0e4ca0cf47 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join16.q.out @@ -0,0 +1,91 @@ +PREHOOK: query: EXPLAIN SELECT subq.key, tab.value FROM (select a.key, a.value from src a where a.key > 10 ) subq JOIN src tab ON (subq.key = tab.key and subq.key > 20 and subq.value = tab.value) where tab.value < 200 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT subq.key, tab.value FROM (select a.key, a.value from src a where a.key > 10 ) subq JOIN src tab ON (subq.key = tab.key and subq.key > 20 and subq.value = tab.value) where tab.value < 200 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: tab + filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(value) < 200.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col1 (type: string) + 1 _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/join17.q.out ql/src/test/results/clientpositive/llap/join17.q.out similarity index 70% rename from ql/src/test/results/clientpositive/join17.q.out rename to ql/src/test/results/clientpositive/llap/join17.q.out index 8d28f35f4e..88f4bf6d71 100644 --- ql/src/test/results/clientpositive/join17.q.out +++ ql/src/test/results/clientpositive/llap/join17.q.out @@ -27,193 +27,288 @@ FROM `default`.`src` WHERE `key` IS NOT NULL) AS `t2` ON `t0`.`key` = `t2`.`key` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: string) - auto parallelism: false - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - Path -> Partition: + Path -> Partition: #### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [$hdt$_0:src1, $hdt$_1:src2] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src1] + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE + Path -> Partition: #### A masked pattern was here #### - table: + Partition + base file name: src input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","key2":"true","value1":"true","value2":"true"}} + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , - columns key1,value1,key2,value2 - columns.comments - columns.types int:string:int:string + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.dest1_n121 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n121 { i32 key1, string value1, i32 key2, string value2} + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n121 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) - outputColumnNames: key1, value1, key2, value2 - Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src2] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","key2":"true","value1":"true","value2":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key1,value1,key2,value2 + columns.comments + columns.types int:string:int:string +#### A masked pattern was here #### + name default.dest1_n121 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n121 { i32 key1, string value1, i32 key2, string value2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n121 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) + outputColumnNames: key1, value1, key2, value2 + Statistics: Num rows: 791 Data size: 150290 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key1, 'hll'), compute_stats(value1, 'hll'), compute_stats(key2, 'hll'), compute_stats(value2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key1, 'hll'), compute_stats(value1, 'hll'), compute_stats(key2, 'hll'), compute_stats(value2, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator - bucketingVersion: 1 + bucketingVersion: 2 compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - column.name.delimiter , + bucketing_version -1 columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct + columns.types struct:struct:struct:struct escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false + Stage: Stage-2 + Dependency Collection + Stage: Stage-0 Move Operator tables: @@ -243,7 +338,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n121 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -253,80 +348,6 @@ STAGE PLANS: Table: default.dest1_n121 Is Table Level Stats: true - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types struct,struct,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - PREHOOK: query: FROM src src1 JOIN src src2 ON (src1.key = src2.key) INSERT OVERWRITE TABLE dest1_n121 SELECT src1.*, src2.* PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/join18.q.out ql/src/test/results/clientpositive/llap/join18.q.out similarity index 59% rename from ql/src/test/results/clientpositive/join18.q.out rename to ql/src/test/results/clientpositive/llap/join18.q.out index 85af7a6a72..cc0f004a27 100644 --- ql/src/test/results/clientpositive/join18.q.out +++ ql/src/test/results/clientpositive/llap/join18.q.out @@ -32,26 +32,73 @@ POSTHOOK: Input: default@src1 #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 5 <- Map 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(value) + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string), value (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 2100 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 2100 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - aggregations: count(value) - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator @@ -61,94 +108,45 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 274 Data size: 2457 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 274 Data size: 2457 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src2 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 262 Data size: 1323 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 262 Data size: 1323 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT value) - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 2196 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 2196 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 2100 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col1) + keys: _col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/join18_multi_distinct.q.out ql/src/test/results/clientpositive/llap/join18_multi_distinct.q.out similarity index 61% rename from ql/src/test/results/clientpositive/join18_multi_distinct.q.out rename to ql/src/test/results/clientpositive/llap/join18_multi_distinct.q.out index 9f7c063c8d..75ef223e47 100644 --- ql/src/test/results/clientpositive/join18_multi_distinct.q.out +++ ql/src/test/results/clientpositive/llap/join18_multi_distinct.q.out @@ -34,26 +34,73 @@ POSTHOOK: Input: default@src1 #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-3 - Stage-3 is a root stage - Stage-0 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 5 <- Map 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(value) + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - aggregations: count(value) - keys: key (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator @@ -63,94 +110,49 @@ STAGE PLANS: Map-reduce partition columns: _col0 (type: string) Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint), _col2 (type: bigint) - Reduce Operator Tree: - Join Operator - condition map: - Full Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 274 Data size: 2561 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 274 Data size: 2561 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src2 - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: key, value - Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 274 Data size: 2561 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 274 Data size: 2561 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - aggregations: count(DISTINCT value), count(DISTINCT key) - keys: key (type: string), value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 12 Data size: 2292 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 12 Data size: 2292 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: count(DISTINCT KEY._col1:0._col0), count(DISTINCT KEY._col1:1._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: CASE WHEN (((_col2 = 0L) and _col1 is not null)) THEN (1) ELSE (null) END (type: int), CASE WHEN (((_col2 = 1L) and _col0 is not null)) THEN (1) ELSE (null) END (type: int), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col0), count(_col1) + keys: _col2 (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint), _col2 (type: bigint) Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join19.q.out ql/src/test/results/clientpositive/llap/join19.q.out new file mode 100644 index 0000000000..0cc787d27c --- /dev/null +++ ql/src/test/results/clientpositive/llap/join19.q.out @@ -0,0 +1,339 @@ +PREHOOK: query: CREATE TABLE triples (foo string, subject string, predicate string, object string, foo2 string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@triples +POSTHOOK: query: CREATE TABLE triples (foo string, subject string, predicate string, object string, foo2 string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@triples +PREHOOK: query: EXPLAIN +SELECT t11.subject, t22.object , t33.subject , t55.object, t66.object +FROM +( +SELECT t1.subject +FROM triples t1 +WHERE +t1.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' +AND +t1.object='http://ontos/OntosMiner/Common.English/ontology#Citation' +) t11 +JOIN +( +SELECT t2.subject , t2.object +FROM triples t2 +WHERE +t2.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' +) t22 +ON (t11.subject=t22.subject) +JOIN +( +SELECT t3.subject , t3.object +FROM triples t3 +WHERE +t3.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_from' + +) t33 +ON (t11.subject=t33.object) +JOIN +( +SELECT t4.subject +FROM triples t4 +WHERE +t4.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' +AND +t4.object='http://ontos/OntosMiner/Common.English/ontology#Author' + +) t44 +ON (t44.subject=t33.subject) +JOIN +( +SELECT t5.subject, t5.object +FROM triples t5 +WHERE +t5.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_to' +) t55 +ON (t55.subject=t44.subject) +JOIN +( +SELECT t6.subject, t6.object +FROM triples t6 +WHERE +t6.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' +) t66 +ON (t66.subject=t55.object) +PREHOOK: type: QUERY +PREHOOK: Input: default@triples +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT t11.subject, t22.object , t33.subject , t55.object, t66.object +FROM +( +SELECT t1.subject +FROM triples t1 +WHERE +t1.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' +AND +t1.object='http://ontos/OntosMiner/Common.English/ontology#Citation' +) t11 +JOIN +( +SELECT t2.subject , t2.object +FROM triples t2 +WHERE +t2.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' +) t22 +ON (t11.subject=t22.subject) +JOIN +( +SELECT t3.subject , t3.object +FROM triples t3 +WHERE +t3.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_from' + +) t33 +ON (t11.subject=t33.object) +JOIN +( +SELECT t4.subject +FROM triples t4 +WHERE +t4.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL' +AND +t4.object='http://ontos/OntosMiner/Common.English/ontology#Author' + +) t44 +ON (t44.subject=t33.subject) +JOIN +( +SELECT t5.subject, t5.object +FROM triples t5 +WHERE +t5.predicate='http://www.ontosearch.com/2007/12/ontosofa-ns#_to' +) t55 +ON (t55.subject=t44.subject) +JOIN +( +SELECT t6.subject, t6.object +FROM triples t6 +WHERE +t6.predicate='http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL' +) t66 +ON (t66.subject=t55.object) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@triples +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 5 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE) + Reducer 6 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t5 + filterExpr: (((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') and subject is not null and object is not null) or ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') and object is not null and subject is not null) or ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null)) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_to') and subject is not null and object is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string), object (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string) + Filter Operator + predicate: ((predicate = 'http://www.ontosearch.com/2007/12/ontosofa-ns#_from') and object is not null and subject is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string), object (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Filter Operator + predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string), object (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: t6 + filterExpr: (((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) or ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author') and subject is not null) or ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation') and subject is not null)) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__LABEL_REL') and subject is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string), object (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Filter Operator + predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Author') and subject is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((predicate = 'http://sofa.semanticweb.org/sofa/v1.0/system#__INSTANCEOF_REL') and (object = 'http://ontos/OntosMiner/Common.English/ontology#Citation') and subject is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: subject (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 552 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col2 (type: string) + outputColumnNames: _col1, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 667 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col5 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col5 (type: string) + Statistics: Num rows: 1 Data size: 667 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col3 (type: string), _col4 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col5 (type: string) + 1 _col2 (type: string) + outputColumnNames: _col1, _col3, _col4, _col8, _col9 + Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col9 (type: string), _col8 (type: string), _col4 (type: string), _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 733 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1 Data size: 607 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/join2.q.out ql/src/test/results/clientpositive/llap/join2.q.out similarity index 58% rename from ql/src/test/results/clientpositive/join2.q.out rename to ql/src/test/results/clientpositive/llap/join2.q.out index cac1cc9d53..b1aaed78d3 100644 --- ql/src/test/results/clientpositive/join2.q.out +++ ql/src/test/results/clientpositive/llap/join2.q.out @@ -26,110 +26,123 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: double) - TableScan - alias: src2 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 49500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 49500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 791 Data size: 81473 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) + key expressions: (_col1 + _col3) (type: double) null sort order: z sort order: + - Map-reduce partition columns: _col0 (type: string) + Map-reduce partition columns: (_col1 + _col3) (type: double) + Statistics: Num rows: 791 Data size: 81473 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 (_col1 + _col3) (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col4 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col4 (type: string) + outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: double) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 791 Data size: 81473 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j2_n2 Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: (_col1 + _col3) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: (_col1 + _col3) (type: double) - Statistics: Num rows: 791 Data size: 81473 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - alias: src3 - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: value (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 49500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 49500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 (_col1 + _col3) (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col4 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col4 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j2_n2 + Dependency Collection Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/join20.q.out ql/src/test/results/clientpositive/llap/join20.q.out similarity index 75% rename from ql/src/test/results/clientpositive/join20.q.out rename to ql/src/test/results/clientpositive/llap/join20.q.out index 02f71ddae0..ac4fbc56db 100644 --- ql/src/test/results/clientpositive/join20.q.out +++ ql/src/test/results/clientpositive/llap/join20.q.out @@ -12,132 +12,123 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) null sort order: z sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter predicates: + 0 + 1 {VALUE._col1} + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: src3 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter predicates: - 0 - 1 {VALUE._col1} - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) - null sort order: zzzzzz - sort order: ++++++ - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -717,132 +708,123 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) null sort order: z sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter predicates: + 0 + 1 {VALUE._col1} + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: src3 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter predicates: - 0 - 1 {VALUE._col1} - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) - null sort order: zzzzzz - sort order: ++++++ - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/join21.q.out ql/src/test/results/clientpositive/llap/join21.q.out similarity index 74% rename from ql/src/test/results/clientpositive/join21.q.out rename to ql/src/test/results/clientpositive/llap/join21.q.out index 58663573a9..c213412e48 100644 --- ql/src/test/results/clientpositive/join21.q.out +++ ql/src/test/results/clientpositive/llap/join21.q.out @@ -10,130 +10,122 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 10.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 0 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) + key expressions: _col2 (type: string) null sort order: z sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 0 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter predicates: + 0 + 1 {VALUE._col1} + keys: + 0 _col2 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string) - TableScan - alias: src3 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 10.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter predicates: - 0 - 1 {VALUE._col1} - keys: - 0 _col2 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) - null sort order: zzzzzz - sort order: ++++++ - Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 501 Data size: 89890 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join22.q.out ql/src/test/results/clientpositive/llap/join22.q.out new file mode 100644 index 0000000000..cfe397d48c --- /dev/null +++ ql/src/test/results/clientpositive/llap/join22.q.out @@ -0,0 +1,119 @@ +PREHOOK: query: explain +SELECT src5.src1_value FROM (SELECT src3.*, src4.value as src4_value, src4.key as src4_key FROM src src4 JOIN (SELECT src2.*, src1.key as src1_key, src1.value as src1_value FROM src src1 JOIN src src2 ON src1.key = src2.key) src3 ON src3.src1_key = src4.key) src5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT src5.src1_value FROM (SELECT src3.*, src4.value as src4_value, src4.key as src4_key FROM src src4 JOIN (SELECT src2.*, src1.key as src1_key, src1.value as src1_value FROM src src1 JOIN src src2 ON src1.key = src2.key) src3 ON src3.src1_key = src4.key) src5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src4 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1 + Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1251 Data size: 113841 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join23.q.out ql/src/test/results/clientpositive/llap/join23.q.out new file mode 100644 index 0000000000..35a2233267 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join23.q.out @@ -0,0 +1,210 @@ +Warning: Shuffle Join MERGEJOIN[15][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 27556 Data size: 9809936 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[15][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM src src1 JOIN src src2 WHERE src1.key < 10 and src2.key < 10 SORT BY src1.key, src1.value, src2.key, src2.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 0 val_0 +0 val_0 2 val_2 +0 val_0 2 val_2 +0 val_0 2 val_2 +0 val_0 4 val_4 +0 val_0 4 val_4 +0 val_0 4 val_4 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 5 val_5 +0 val_0 8 val_8 +0 val_0 8 val_8 +0 val_0 8 val_8 +0 val_0 9 val_9 +0 val_0 9 val_9 +0 val_0 9 val_9 +2 val_2 0 val_0 +2 val_2 0 val_0 +2 val_2 0 val_0 +2 val_2 2 val_2 +2 val_2 4 val_4 +2 val_2 5 val_5 +2 val_2 5 val_5 +2 val_2 5 val_5 +2 val_2 8 val_8 +2 val_2 9 val_9 +4 val_4 0 val_0 +4 val_4 0 val_0 +4 val_4 0 val_0 +4 val_4 2 val_2 +4 val_4 4 val_4 +4 val_4 5 val_5 +4 val_4 5 val_5 +4 val_4 5 val_5 +4 val_4 8 val_8 +4 val_4 9 val_9 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 0 val_0 +5 val_5 2 val_2 +5 val_5 2 val_2 +5 val_5 2 val_2 +5 val_5 4 val_4 +5 val_5 4 val_4 +5 val_5 4 val_4 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 5 val_5 +5 val_5 8 val_8 +5 val_5 8 val_8 +5 val_5 8 val_8 +5 val_5 9 val_9 +5 val_5 9 val_9 +5 val_5 9 val_9 +8 val_8 0 val_0 +8 val_8 0 val_0 +8 val_8 0 val_0 +8 val_8 2 val_2 +8 val_8 4 val_4 +8 val_8 5 val_5 +8 val_8 5 val_5 +8 val_8 5 val_5 +8 val_8 8 val_8 +8 val_8 9 val_9 +9 val_9 0 val_0 +9 val_9 0 val_0 +9 val_9 0 val_0 +9 val_9 2 val_2 +9 val_9 4 val_4 +9 val_9 5 val_5 +9 val_9 5 val_5 +9 val_9 5 val_5 +9 val_9 8 val_8 +9 val_9 9 val_9 diff --git ql/src/test/results/clientpositive/llap/join25.q.out ql/src/test/results/clientpositive/llap/join25.q.out new file mode 100644 index 0000000000..8d72e5a999 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join25.q.out @@ -0,0 +1,212 @@ +PREHOOK: query: CREATE TABLE dest_j1_n18(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n18 +POSTHOOK: query: CREATE TABLE dest_j1_n18(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n18 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n18 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n18 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n18 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n18 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 2 <- Map 1 (BROADCAST_EDGE) + Reducer 3 <- Map 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + input vertices: + 0 Map 1 + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n18 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.974359 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n18 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value, val2 + Column Types: int, string, string + Table: default.dest_j1_n18 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n18 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n18 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n18 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n18 +POSTHOOK: Lineage: dest_j1_n18.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n18.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n18.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n18 x +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n18 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n18 x +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n18 +#### A masked pattern was here #### +128 val_128 +128 val_128 +128 val_128 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +224 val_224 +224 val_224 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +369 val_369 +369 val_369 +369 val_369 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 diff --git ql/src/test/results/clientpositive/llap/join26.q.out ql/src/test/results/clientpositive/llap/join26.q.out new file mode 100644 index 0000000000..724ec9e371 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join26.q.out @@ -0,0 +1,583 @@ +PREHOOK: query: CREATE TABLE dest_j1_n10(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n10 +POSTHOOK: query: CREATE TABLE dest_j1_n10(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n10 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n10 +SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n10 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n10 +SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n10 +OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` +FROM (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `key` IS NOT NULL) AS `t0` +INNER JOIN ((SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` IS NOT NULL) AS `t2` +INNER JOIN (SELECT `key` +FROM `default`.`src1` +WHERE `key` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`key` = `t4`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) + Map 3 <- Map 1 (BROADCAST_EDGE) + Reducer 4 <- Map 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 2 => 25 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2 + input vertices: + 1 Map 2 + Position of Big Table: 0 + Statistics: Num rows: 39 Data size: 6903 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col2 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 39 Data size: 6903 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [y] + Map 2 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src1 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src1 + name: default.src1 + Truncated Path -> Alias: + /src1 [x] + Map 3 + Map Operator Tree: + TableScan + alias: z + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 1 => 39 + keys: + 0 _col2 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col4 + input vertices: + 0 Map 1 + Position of Big Table: 1 + Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n10 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n10 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n10 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 61 Data size: 16348 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.9836066 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [z] + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n10 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n10 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n10 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, val2 + Column Types: string, string, string + Table: default.dest_j1_n10 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n10 +SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n10 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n10 +SELECT /*+ MAPJOIN(x,y) */ x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.key = z.key and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n10 +POSTHOOK: Lineage: dest_j1_n10.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n10.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n10.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n10 x +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n10 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n10 x +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n10 +#### A masked pattern was here #### +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +128 val_128 val_128 +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +224 val_224 val_224 +224 val_224 val_224 +224 val_224 val_224 +224 val_224 val_224 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +369 val_369 val_369 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 diff --git ql/src/test/results/clientpositive/llap/join27.q.out ql/src/test/results/clientpositive/llap/join27.q.out new file mode 100644 index 0000000000..4328cd6bd6 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join27.q.out @@ -0,0 +1,216 @@ +PREHOOK: query: CREATE TABLE dest_j1_n2(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n2 +POSTHOOK: query: CREATE TABLE dest_j1_n2(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n2 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n2 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.value = y.value) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n2 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n2 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.value = y.value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n2 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 2 <- Map 1 (BROADCAST_EDGE) + Reducer 3 <- Map 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: y + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + input vertices: + 0 Map 1 + Statistics: Num rows: 40 Data size: 10640 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n2 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 40 Data size: 7360 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.975 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value, val2 + Column Types: int, string, string + Table: default.dest_j1_n2 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n2 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.value = y.value) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n2 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n2 +SELECT /*+ MAPJOIN(x) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.value = y.value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n2 +POSTHOOK: Lineage: dest_j1_n2.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n2.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n2.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n2 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n2 +#### A masked pattern was here #### +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 +NULL val_165 val_165 +NULL val_165 val_165 +NULL val_193 val_193 +NULL val_193 val_193 +NULL val_193 val_193 +NULL val_265 val_265 +NULL val_265 val_265 +NULL val_27 val_27 +NULL val_409 val_409 +NULL val_409 val_409 +NULL val_409 val_409 +NULL val_484 val_484 diff --git ql/src/test/results/clientpositive/llap/join28.q.out ql/src/test/results/clientpositive/llap/join28.q.out new file mode 100644 index 0000000000..485158dc21 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join28.q.out @@ -0,0 +1,332 @@ +PREHOOK: query: CREATE TABLE dest_j1_n11(key STRING, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n11 +POSTHOOK: query: CREATE TABLE dest_j1_n11(key STRING, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n11 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n11 +SELECT subq.key1, z.value +FROM +(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 + FROM src1 x JOIN src y ON (x.key = y.key)) subq + JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n11 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n11 +SELECT subq.key1, z.value +FROM +(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 + FROM src1 x JOIN src y ON (x.key = y.key)) subq + JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n11 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) + Map 3 <- Map 1 (BROADCAST_EDGE) + Reducer 4 <- Map 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1 + input vertices: + 1 Map 2 + Statistics: Num rows: 39 Data size: 3354 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 39 Data size: 3354 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: z + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col3 + input vertices: + 0 Map 1 + Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n11 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 61 Data size: 10797 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.9836066 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n11 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.dest_j1_n11 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n11 +SELECT subq.key1, z.value +FROM +(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 + FROM src1 x JOIN src y ON (x.key = y.key)) subq + JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n11 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n11 +SELECT subq.key1, z.value +FROM +(SELECT x.key as key1, x.value as value1, y.key as key2, y.value as value2 + FROM src1 x JOIN src y ON (x.key = y.key)) subq + JOIN srcpart z ON (subq.key1 = z.key and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n11 +POSTHOOK: Lineage: dest_j1_n11.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n11.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n11 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n11 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n11 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n11 +#### A masked pattern was here #### +128 val_128 +128 val_128 +128 val_128 +128 val_128 +128 val_128 +128 val_128 +128 val_128 +128 val_128 +128 val_128 +146 val_146 +146 val_146 +146 val_146 +146 val_146 +150 val_150 +213 val_213 +213 val_213 +213 val_213 +213 val_213 +224 val_224 +224 val_224 +224 val_224 +224 val_224 +238 val_238 +238 val_238 +238 val_238 +238 val_238 +255 val_255 +255 val_255 +255 val_255 +255 val_255 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +273 val_273 +278 val_278 +278 val_278 +278 val_278 +278 val_278 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +311 val_311 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +369 val_369 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +401 val_401 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +406 val_406 +66 val_66 +98 val_98 +98 val_98 +98 val_98 +98 val_98 diff --git ql/src/test/results/clientpositive/llap/join29.q.out ql/src/test/results/clientpositive/llap/join29.q.out new file mode 100644 index 0000000000..b2df356f2e --- /dev/null +++ ql/src/test/results/clientpositive/llap/join29.q.out @@ -0,0 +1,233 @@ +PREHOOK: query: CREATE TABLE dest_j1_n6(key STRING, cnt1 INT, cnt2 INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n6 +POSTHOOK: query: CREATE TABLE dest_j1_n6(key STRING, cnt1 INT, cnt2 INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n6 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n6 +SELECT subq1.key, subq1.cnt, subq2.cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n6 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n6 +SELECT subq1.key, subq1.cnt, subq2.cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n6 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 2 (BROADCAST_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.52 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 23750 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + input vertices: + 0 Reducer 2 + Statistics: Num rows: 12 Data size: 1224 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int), UDFToInteger(_col3) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n6 + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: int) + outputColumnNames: key, cnt1, cnt2 + Statistics: Num rows: 12 Data size: 1128 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt1, 'hll'), compute_stats(cnt2, 'hll') + minReductionHashAggr: 0.9166667 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1288 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n6 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt1, cnt2 + Column Types: string, int, int + Table: default.dest_j1_n6 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n6 +SELECT subq1.key, subq1.cnt, subq2.cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n6 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n6 +SELECT subq1.key, subq1.cnt, subq2.cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n6 +POSTHOOK: Lineage: dest_j1_n6.cnt1 EXPRESSION [(src1)x.null, ] +POSTHOOK: Lineage: dest_j1_n6.cnt2 EXPRESSION [(src)y.null, ] +POSTHOOK: Lineage: dest_j1_n6.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n6 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n6 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n6 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n6 +#### A masked pattern was here #### +128 1 3 +146 1 2 +150 1 1 +213 1 2 +224 1 2 +238 1 2 +255 1 2 +273 1 3 +278 1 2 +311 1 3 +369 1 3 +401 1 5 +406 1 4 +66 1 1 +98 1 2 diff --git ql/src/test/results/clientpositive/join3.q.out ql/src/test/results/clientpositive/llap/join3.q.out similarity index 83% rename from ql/src/test/results/clientpositive/join3.q.out rename to ql/src/test/results/clientpositive/llap/join3.q.out index 787506f65e..237c0a3690 100644 --- ql/src/test/results/clientpositive/join3.q.out +++ ql/src/test/results/clientpositive/llap/join3.q.out @@ -22,128 +22,155 @@ STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0, Stage-4 - Stage-4 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 791 Data size: 68817 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 791 Data size: 68817 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src3 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: src3 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 791 Data size: 68817 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) null sort order: z sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col3 - Statistics: Num rows: 1251 Data size: 222678 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n46 - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 791 Data size: 68817 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 1251 Data size: 222678 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n46 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 1251 Data size: 118845 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -163,30 +190,6 @@ STAGE PLANS: Column Types: int, string Table: default.dest1_n46 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM src src1 JOIN src src2 ON (src1.key = src2.key) JOIN src src3 ON (src1.key = src3.key) INSERT OVERWRITE TABLE dest1_n46 SELECT src1.key, src3.value PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/join30.q.out ql/src/test/results/clientpositive/llap/join30.q.out new file mode 100644 index 0000000000..0a79ce5848 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join30.q.out @@ -0,0 +1,208 @@ +PREHOOK: query: CREATE TABLE dest_j1_n0(key INT, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n0 +POSTHOOK: query: CREATE TABLE dest_j1_n0(key INT, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n0 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n0 +SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n0 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n0 +SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n0 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 2 <- Map 1 (BROADCAST_EDGE) + Reducer 3 <- Map 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0 + input vertices: + 0 Map 1 + Statistics: Num rows: 39 Data size: 3354 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: string) + minReductionHashAggr: 0.5897436 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 16 Data size: 1504 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n0 + Select Operator + expressions: _col0 (type: int), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 16 Data size: 128 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.9375 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 848 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n0 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: int, int + Table: default.dest_j1_n0 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n0 +SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n0 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n0 +SELECT /*+ MAPJOIN(x) */ x.key, count(1) FROM src1 x JOIN src y ON (x.key = y.key) group by x.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n0 +POSTHOOK: Lineage: dest_j1_n0.cnt EXPRESSION [(src1)x.null, (src)y.null, ] +POSTHOOK: Lineage: dest_j1_n0.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n0 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n0 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n0 +#### A masked pattern was here #### +128 3 +146 2 +150 1 +213 2 +224 2 +238 2 +255 2 +273 3 +278 2 +311 3 +369 3 +401 5 +406 4 +66 1 +98 2 diff --git ql/src/test/results/clientpositive/llap/join31.q.out ql/src/test/results/clientpositive/llap/join31.q.out new file mode 100644 index 0000000000..20914a471d --- /dev/null +++ ql/src/test/results/clientpositive/llap/join31.q.out @@ -0,0 +1,242 @@ +PREHOOK: query: CREATE TABLE dest_j1_n22(key STRING, cnt INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n22 +POSTHOOK: query: CREATE TABLE dest_j1_n22(key STRING, cnt INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n22 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n22 +SELECT subq1.key, count(1) as cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +group by subq1.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n22 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n22 +SELECT subq1.key, count(1) as cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +group by subq1.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n22 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (BROADCAST_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: string) + minReductionHashAggr: 0.52 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 21750 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0 + input vertices: + 1 Map 5 + Statistics: Num rows: 12 Data size: 1032 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 564 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n22 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, cnt + Statistics: Num rows: 6 Data size: 540 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(cnt, 'hll') + minReductionHashAggr: 0.8333333 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n22 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, cnt + Column Types: string, int + Table: default.dest_j1_n22 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n22 +SELECT subq1.key, count(1) as cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +group by subq1.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n22 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n22 +SELECT subq1.key, count(1) as cnt +FROM (select x.key, count(1) as cnt from src1 x group by x.key) subq1 JOIN + (select y.key, count(1) as cnt from src y group by y.key) subq2 ON (subq1.key = subq2.key) +group by subq1.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n22 +POSTHOOK: Lineage: dest_j1_n22.cnt EXPRESSION [(src1)x.null, ] +POSTHOOK: Lineage: dest_j1_n22.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n22 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n22 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n22 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n22 +#### A masked pattern was here #### +128 1 +146 1 +150 1 +213 1 +224 1 +238 1 +255 1 +273 1 +278 1 +311 1 +369 1 +401 1 +406 1 +66 1 +98 1 diff --git ql/src/test/results/clientpositive/llap/join32.q.out ql/src/test/results/clientpositive/llap/join32.q.out new file mode 100644 index 0000000000..24b3e118dd --- /dev/null +++ ql/src/test/results/clientpositive/llap/join32.q.out @@ -0,0 +1,562 @@ +PREHOOK: query: CREATE TABLE dest_j1_n12(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n12 +POSTHOOK: query: CREATE TABLE dest_j1_n12(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n12 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n12 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n12 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n12 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n12 +OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` +FROM (SELECT `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `value` IS NOT NULL) AS `t0` +INNER JOIN ((SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` IS NOT NULL) AS `t2` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src1` +WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`value` = `t4`.`value` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) + Map 3 <- Map 1 (BROADCAST_EDGE) + Reducer 4 <- Map 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 2 => 25 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 2 + Position of Big Table: 0 + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col3 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col3 (type: string) + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string), _col2 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [y] + Map 2 + Map Operator Tree: + TableScan + alias: x + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src1 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src1 + name: default.src1 + Truncated Path -> Alias: + /src1 [x] + Map 3 + Map Operator Tree: + TableScan + alias: z + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 1 => 39 + keys: + 0 _col3 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col4 + input vertices: + 0 Map 1 + Position of Big Table: 1 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n12 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n12 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n12 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.984127 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [z] + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n12 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n12 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n12 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, val2 + Column Types: string, string, string + Table: default.dest_j1_n12 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n12 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n12 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n12 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n12 +POSTHOOK: Lineage: dest_j1_n12.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n12.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n12.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n12 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n12 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n12 +#### A masked pattern was here #### +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 diff --git ql/src/test/results/clientpositive/llap/join33.q.out ql/src/test/results/clientpositive/llap/join33.q.out new file mode 100644 index 0000000000..530e4a5e30 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join33.q.out @@ -0,0 +1,562 @@ +PREHOOK: query: CREATE TABLE dest_j1_n7(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n7 +POSTHOOK: query: CREATE TABLE dest_j1_n7(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n7 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n7 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n7 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n7 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n7 +OPTIMIZED SQL: SELECT `t4`.`key`, `t0`.`value`, `t2`.`value` AS `value1` +FROM (SELECT `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' AND `hr` = 11 AND `value` IS NOT NULL) AS `t0` +INNER JOIN ((SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` IS NOT NULL) AS `t2` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src1` +WHERE `key` IS NOT NULL AND `value` IS NOT NULL) AS `t4` ON `t2`.`key` = `t4`.`key`) ON `t0`.`value` = `t4`.`value` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) + Map 3 <- Map 1 (BROADCAST_EDGE) + Reducer 4 <- Map 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 2 => 25 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 2 + Position of Big Table: 0 + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col3 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col3 (type: string) + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string), _col2 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [y] + Map 2 + Map Operator Tree: + TableScan + alias: x + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src1 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src1 + name: default.src1 + Truncated Path -> Alias: + /src1 [x] + Map 3 + Map Operator Tree: + TableScan + alias: z + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 1 => 39 + keys: + 0 _col3 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col4 + input vertices: + 0 Map 1 + Position of Big Table: 1 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col4 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n7 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n7 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n7 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 63 Data size: 16884 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.984127 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [z] + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n7 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n7 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n7 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, val2 + Column Types: string, string, string + Table: default.dest_j1_n7 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n7 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Output: default@dest_j1_n7 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n7 +SELECT x.key, z.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +JOIN srcpart z ON (x.value = z.value and z.ds='2008-04-08' and z.hr=11) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Output: default@dest_j1_n7 +POSTHOOK: Lineage: dest_j1_n7.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n7.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n7.value SIMPLE [(srcpart)z.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n7 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n7 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n7 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n7 +#### A masked pattern was here #### +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +213 val_213 val_213 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 +98 val_98 val_98 diff --git ql/src/test/results/clientpositive/llap/join34.q.out ql/src/test/results/clientpositive/llap/join34.q.out new file mode 100644 index 0000000000..0802efe2ac --- /dev/null +++ ql/src/test/results/clientpositive/llap/join34.q.out @@ -0,0 +1,578 @@ +PREHOOK: query: CREATE TABLE dest_j1_n1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n1 +POSTHOOK: query: CREATE TABLE dest_j1_n1(key STRING, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n1 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n1 +SELECT x.key, x.value, subq1.value +FROM +( SELECT x.key as key, x.value as value from src x where x.key < 20 + UNION ALL + SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 +) subq1 +JOIN src1 x ON (x.key = subq1.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n1 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n1 +SELECT x.key, x.value, subq1.value +FROM +( SELECT x.key as key, x.value as value from src x where x.key < 20 + UNION ALL + SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 +) subq1 +JOIN src1 x ON (x.key = subq1.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n1 +OPTIMIZED SQL: SELECT `t5`.`key`, `t5`.`value`, `t3`.`value` AS `value1` +FROM (SELECT `key`, `value` +FROM `default`.`src` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` < 20 +UNION ALL +SELECT `key`, `value` +FROM `default`.`src` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` > 100) AS `t3` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src1` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` IS NOT NULL) AS `t5` ON `t3`.`key` = `t5`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 5 (BROADCAST_EDGE), Union 2 (CONTAINS) + Map 4 <- Map 5 (BROADCAST_EDGE), Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 5 => 23 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 5 + Position of Big Table: 0 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col3 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.9782609 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [x] + Map 4 + Map Operator Tree: + TableScan + alias: x1 + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) + Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 148 Data size: 26344 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 5 => 23 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 5 + Position of Big Table: 0 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col3 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 46 Data size: 12236 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.9782609 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [x1] + Map 5 + Map Operator Tree: + TableScan + alias: x + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Reduce Output Operator + bucketingVersion: -2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src1 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src1 + name: default.src1 + Truncated Path -> Alias: + /src1 [x] + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:string +#### A masked pattern was here #### + name default.dest_j1_n1 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n1 { string key, string value, string val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, val2 + Column Types: string, string, string + Table: default.dest_j1_n1 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n1 +SELECT x.key, x.value, subq1.value +FROM +( SELECT x.key as key, x.value as value from src x where x.key < 20 + UNION ALL + SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 +) subq1 +JOIN src1 x ON (x.key = subq1.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n1 +SELECT x.key, x.value, subq1.value +FROM +( SELECT x.key as key, x.value as value from src x where x.key < 20 + UNION ALL + SELECT x1.key as key, x1.value as value from src x1 where x1.key > 100 +) subq1 +JOIN src1 x ON (x.key = subq1.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n1 +POSTHOOK: Lineage: dest_j1_n1.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n1.val2 EXPRESSION [(src)x.FieldSchema(name:value, type:string, comment:default), (src)x1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n1.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n1 +#### A masked pattern was here #### +128 val_128 +128 val_128 +128 val_128 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +224 val_224 +224 val_224 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +369 val_369 +369 val_369 +369 val_369 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 diff --git ql/src/test/results/clientpositive/llap/join35.q.out ql/src/test/results/clientpositive/llap/join35.q.out new file mode 100644 index 0000000000..fb4a512e0d --- /dev/null +++ ql/src/test/results/clientpositive/llap/join35.q.out @@ -0,0 +1,607 @@ +PREHOOK: query: CREATE TABLE dest_j1_n24(key STRING, value STRING, val2 INT) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n24 +POSTHOOK: query: CREATE TABLE dest_j1_n24(key STRING, value STRING, val2 INT) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n24 +PREHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n24 +SELECT x.key, x.value, subq1.cnt +FROM +( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key + UNION ALL + SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key +) subq1 +JOIN src1 x ON (x.key = subq1.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n24 +POSTHOOK: query: EXPLAIN EXTENDED +INSERT OVERWRITE TABLE dest_j1_n24 +SELECT x.key, x.value, subq1.cnt +FROM +( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key + UNION ALL + SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key +) subq1 +JOIN src1 x ON (x.key = subq1.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n24 +OPTIMIZED SQL: SELECT `t5`.`key`, `t5`.`value`, `t3`.`$f1` AS `cnt` +FROM (SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`src` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` < 20 +GROUP BY `key` +UNION ALL +SELECT `key`, COUNT(*) AS `$f1` +FROM `default`.`src` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` > 100 +GROUP BY `key`) AS `t3` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src1` +WHERE CAST(`key` AS DOUBLE) BETWEEN 20 AND 100 AND `key` IS NOT NULL) AS `t5` ON `t3`.`key` = `t5`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 7 (BROADCAST_EDGE), Union 3 (CONTAINS) + Reducer 4 <- Union 3 (CUSTOM_SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE), Map 7 (BROADCAST_EDGE), Union 3 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 148 Data size: 12876 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [x] + Map 5 + Map Operator Tree: + TableScan + alias: x1 + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and (UDFToDouble(key) > 100.0D)) (type: boolean) + Statistics: Num rows: 148 Data size: 12876 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: key (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [x1] + Map 7 + Map Operator Tree: + TableScan + alias: x + filterExpr: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (UDFToDouble(key) NOT BETWEEN 20.0D AND 100.0D and key is not null) (type: boolean) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Reduce Output Operator + bucketingVersion: -2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 23 Data size: 4025 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src1 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src1 + numFiles 1 + numRows 25 + rawDataSize 191 + serialization.ddl struct src1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 216 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src1 + name: default.src1 + Truncated Path -> Alias: + /src1 [x] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 7 => 23 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 7 + Position of Big Table: 0 + Statistics: Num rows: 46 Data size: 8418 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col3 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:int +#### A masked pattern was here #### + name default.dest_j1_n24 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n24 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, value, val2 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.9782609 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 6 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 74 Data size: 7030 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + Estimated key counts: Map 7 => 23 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + input vertices: + 1 Map 7 + Position of Big Table: 0 + Statistics: Num rows: 46 Data size: 8418 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col3 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:int +#### A masked pattern was here #### + name default.dest_j1_n24 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n24 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int) + outputColumnNames: key, value, val2 + Statistics: Num rows: 46 Data size: 8234 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.9782609 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + auto parallelism: false + Union 3 + Vertex: Union 3 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value,val2 + columns.comments + columns.types string:string:int +#### A masked pattern was here #### + name default.dest_j1_n24 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n24 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value, val2 + Column Types: string, string, int + Table: default.dest_j1_n24 + Is Table Level Stats: true + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n24 +SELECT x.key, x.value, subq1.cnt +FROM +( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key + UNION ALL + SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key +) subq1 +JOIN src1 x ON (x.key = subq1.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n24 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n24 +SELECT x.key, x.value, subq1.cnt +FROM +( SELECT x.key as key, count(1) as cnt from src x where x.key < 20 group by x.key + UNION ALL + SELECT x1.key as key, count(1) as cnt from src x1 where x1.key > 100 group by x1.key +) subq1 +JOIN src1 x ON (x.key = subq1.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n24 +POSTHOOK: Lineage: dest_j1_n24.key SIMPLE [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n24.val2 EXPRESSION [(src)x.null, (src)x1.null, ] +POSTHOOK: Lineage: dest_j1_n24.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n24 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n24 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n24 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n24 +#### A masked pattern was here #### +128 3 +146 val_146 2 +150 val_150 1 +213 val_213 2 +224 2 +238 val_238 2 +255 val_255 2 +273 val_273 3 +278 val_278 2 +311 val_311 3 +369 3 +401 val_401 5 +406 val_406 4 diff --git ql/src/test/results/clientpositive/join36.q.out ql/src/test/results/clientpositive/llap/join36.q.out similarity index 54% rename from ql/src/test/results/clientpositive/join36.q.out rename to ql/src/test/results/clientpositive/llap/join36.q.out index 33f7b72996..8071b04748 100644 --- ql/src/test/results/clientpositive/join36.q.out +++ ql/src/test/results/clientpositive/llap/join36.q.out @@ -63,89 +63,112 @@ POSTHOOK: Input: default@tmp1_n0 POSTHOOK: Input: default@tmp2_n0 POSTHOOK: Output: default@dest_j1_n13 STAGE DEPENDENCIES: - Stage-6 is a root stage - Stage-5 depends on stages: Stage-6 - Stage-0 depends on stages: Stage-5 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-5 + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: - Stage: Stage-6 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:x - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:x - TableScan - alias: x - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), cnt (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: y - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), cnt (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col3 - Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n13 + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 3 (BROADCAST_EDGE) + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) - outputColumnNames: key, value, val2 - Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1272 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work + expressions: key (type: int), cnt (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3 + input vertices: + 1 Map 3 + Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n13 + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + outputColumnNames: key, value, val2 + Statistics: Num rows: 315 Data size: 3780 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1272 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1272 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), cnt (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 309 Data size: 2472 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -157,7 +180,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n13 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -165,30 +188,6 @@ STAGE PLANS: Column Types: int, int, int Table: default.dest_j1_n13 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1272 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n13 SELECT /*+ MAPJOIN(x) */ x.key, x.cnt, y.cnt FROM tmp1_n0 x JOIN tmp2_n0 y ON (x.key = y.key) diff --git ql/src/test/results/clientpositive/llap/join37.q.out ql/src/test/results/clientpositive/llap/join37.q.out new file mode 100644 index 0000000000..3c09084c9d --- /dev/null +++ ql/src/test/results/clientpositive/llap/join37.q.out @@ -0,0 +1,212 @@ +PREHOOK: query: CREATE TABLE dest_j1_n9(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_j1_n9 +POSTHOOK: query: CREATE TABLE dest_j1_n9(key INT, value STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_j1_n9 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n9 +SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n9 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_j1_n9 +SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n9 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 2 <- Map 1 (BROADCAST_EDGE) + Reducer 3 <- Map 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: y + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3 + input vertices: + 0 Map 1 + Statistics: Num rows: 39 Data size: 10374 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n9 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, val2 + Statistics: Num rows: 39 Data size: 7176 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.974359 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1304 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1320 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n9 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value, val2 + Column Types: int, string, string + Table: default.dest_j1_n9 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n9 +SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_j1_n9 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_j1_n9 +SELECT /*+ MAPJOIN(X) */ x.key, x.value, y.value +FROM src1 x JOIN src y ON (x.key = y.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_j1_n9 +POSTHOOK: Lineage: dest_j1_n9.key EXPRESSION [(src1)x.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n9.val2 SIMPLE [(src)y.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest_j1_n9.value SIMPLE [(src1)x.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: select * from dest_j1_n9 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest_j1_n9 +#### A masked pattern was here #### +POSTHOOK: query: select * from dest_j1_n9 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest_j1_n9 +#### A masked pattern was here #### +128 val_128 +128 val_128 +128 val_128 +146 val_146 val_146 +146 val_146 val_146 +150 val_150 val_150 +213 val_213 val_213 +213 val_213 val_213 +224 val_224 +224 val_224 +238 val_238 val_238 +238 val_238 val_238 +255 val_255 val_255 +255 val_255 val_255 +273 val_273 val_273 +273 val_273 val_273 +273 val_273 val_273 +278 val_278 val_278 +278 val_278 val_278 +311 val_311 val_311 +311 val_311 val_311 +311 val_311 val_311 +369 val_369 +369 val_369 +369 val_369 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +401 val_401 val_401 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +406 val_406 val_406 +66 val_66 val_66 +98 val_98 val_98 +98 val_98 val_98 diff --git ql/src/test/results/clientpositive/llap/join38.q.out ql/src/test/results/clientpositive/llap/join38.q.out new file mode 100644 index 0000000000..4c258445cd --- /dev/null +++ ql/src/test/results/clientpositive/llap/join38.q.out @@ -0,0 +1,170 @@ +PREHOOK: query: create table tmp_n1(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tmp_n1 +POSTHOOK: query: create table tmp_n1(col0 string, col1 string,col2 string,col3 string,col4 string,col5 string,col6 string,col7 string,col8 string,col9 string,col10 string,col11 string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tmp_n1 +PREHOOK: query: insert overwrite table tmp_n1 select key, cast(key + 1 as int), key +2, key+3, key+4, cast(key+5 as int), key+6, key+7, key+8, key+9, key+10, cast(key+11 as int) from src where key = 100 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tmp_n1 +POSTHOOK: query: insert overwrite table tmp_n1 select key, cast(key + 1 as int), key +2, key+3, key+4, cast(key+5 as int), key+6, key+7, key+8, key+9, key+10, cast(key+11 as int) from src where key = 100 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tmp_n1 +POSTHOOK: Lineage: tmp_n1.col0 SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: tmp_n1.col1 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col10 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col11 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col2 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col3 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col4 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col5 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col6 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col7 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col8 EXPRESSION [] +POSTHOOK: Lineage: tmp_n1.col9 EXPRESSION [] +PREHOOK: query: select * from tmp_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +POSTHOOK: query: select * from tmp_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +100 101 102.0 103.0 104.0 105 106.0 107.0 108.0 109.0 110.0 111 +100 101 102.0 103.0 104.0 105 106.0 107.0 108.0 109.0 110.0 111 +PREHOOK: query: explain +FROM src a JOIN tmp_n1 b ON (a.key = b.col11) +SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count +where b.col11 = 111 +group by a.value, b.col5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +POSTHOOK: query: explain +FROM src a JOIN tmp_n1 b ON (a.key = b.col11) +SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count +where b.col11 = 111 +group by a.value, b.col5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 3 (BROADCAST_EDGE) + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: (UDFToDouble(key) = 111.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) = 111.0D) (type: boolean) + Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 44500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col1, _col2 + input vertices: + 1 Map 3 + Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + keys: _col1 (type: string), _col2 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: (UDFToDouble(col11) = 111.0D) (type: boolean) + Statistics: Num rows: 2 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(col11) = 111.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: col5 (type: string), col11 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM src a JOIN tmp_n1 b ON (a.key = b.col11) +SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count +where b.col11 = 111 +group by a.value, b.col5 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +POSTHOOK: query: FROM src a JOIN tmp_n1 b ON (a.key = b.col11) +SELECT /*+ MAPJOIN(a) */ a.value, b.col5, count(1) as count +where b.col11 = 111 +group by a.value, b.col5 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@tmp_n1 +#### A masked pattern was here #### +val_111 105 2 diff --git ql/src/test/results/clientpositive/join39.q.out ql/src/test/results/clientpositive/llap/join39.q.out similarity index 73% rename from ql/src/test/results/clientpositive/join39.q.out rename to ql/src/test/results/clientpositive/llap/join39.q.out index ff8d2980e4..a2bc64d1b8 100644 --- ql/src/test/results/clientpositive/join39.q.out +++ ql/src/test/results/clientpositive/llap/join39.q.out @@ -21,81 +21,104 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src POSTHOOK: Output: default@dest_j1_n8 STAGE DEPENDENCIES: - Stage-6 is a root stage - Stage-5 depends on stages: Stage-6 - Stage-0 depends on stages: Stage-5 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-5 + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: - Stage: Stage-6 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:src - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:src - TableScan - alias: src - filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - alias: x - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 3 (BROADCAST_EDGE) + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + input vertices: + 1 Map 3 + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_j1_n8 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, key1, val2 + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(key1, 'hll'), compute_stats(val2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest_j1_n8 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, key1, val2 - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll'), compute_stats(key1, 'hll'), compute_stats(val2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Local Work: - Map Reduce Local Work + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -107,7 +130,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n8 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -115,30 +138,6 @@ STAGE PLANS: Column Types: string, string, string, string Table: default.dest_j1_n8 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: INSERT OVERWRITE TABLE dest_j1_n8 SELECT /*+ MAPJOIN(y) */ x.key, x.value, y.key, y.value FROM src x left outer JOIN (select * from src where key <= 100) y ON (x.key = y.key) diff --git ql/src/test/results/clientpositive/llap/join4.q.out ql/src/test/results/clientpositive/llap/join4.q.out new file mode 100644 index 0000000000..7eae3ae6d7 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join4.q.out @@ -0,0 +1,231 @@ +PREHOOK: query: CREATE TABLE dest1_n72(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n72 +POSTHOOK: query: CREATE TABLE dest1_n72(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n72 +PREHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n72 +POSTHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n72 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n72 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) + outputColumnNames: c1, c2, c3, c4 + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') + minReductionHashAggr: 0.9818182 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n72 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3, c4 + Column Types: int, string, int, string + Table: default.dest1_n72 + +PREHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n72 +POSTHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n72 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n72 +POSTHOOK: Lineage: dest1_n72.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n72.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n72.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n72.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n72.* FROM dest1_n72 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n72 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n72.* FROM dest1_n72 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n72 +#### A masked pattern was here #### +11 val_11 NULL NULL +12 val_12 NULL NULL +12 val_12 NULL NULL +15 val_15 NULL NULL +15 val_15 NULL NULL +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 diff --git ql/src/test/results/clientpositive/join40.q.out ql/src/test/results/clientpositive/llap/join40.q.out similarity index 75% rename from ql/src/test/results/clientpositive/join40.q.out rename to ql/src/test/results/clientpositive/llap/join40.q.out index f33ff5a0d6..7efab037aa 100644 --- ql/src/test/results/clientpositive/join40.q.out +++ ql/src/test/results/clientpositive/llap/join40.q.out @@ -14,56 +14,70 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: x - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -663,63 +677,77 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -1779,132 +1807,123 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) null sort order: z sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter predicates: + 0 + 1 {VALUE._col1} + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: src3 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter predicates: - 0 - 1 {VALUE._col1} - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) - null sort order: zzzzzz - sort order: ++++++ - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -2484,132 +2503,123 @@ POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src3 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Filter Operator + predicate: (UDFToDouble(key) < 10.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator key expressions: _col0 (type: string) null sort order: z sort order: + Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src2 - filterExpr: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 10.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter predicates: + 0 + 1 {VALUE._col1} + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 166 Data size: 59096 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - alias: src3 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 20.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - Reduce Operator Tree: - Join Operator - condition map: - Right Outer Join 0 to 1 - filter predicates: - 0 - 1 {VALUE._col1} - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) - null sort order: zzzzzz - sort order: ++++++ - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 229264 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -3188,63 +3198,67 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-0 depends on stages: Stage-3 + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-4 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_1:src - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_1:src - TableScan - alias: src - filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: x - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Local Work: - Map Reduce Local Work + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: x + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + input vertices: + 1 Map 2 + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 762 Data size: 182450 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) <= 100.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs Stage: Stage-0 Fetch Operator @@ -3839,83 +3853,89 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-5 is a root stage - Stage-2 depends on stages: Stage-5 - Stage-0 depends on stages: Stage-2 + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-5 - Map Reduce Local Work - Alias -> Map Local Tables: - $hdt$_0:a - Fetch Operator - limit: -1 - Alias -> Map Local Operator Tree: - $hdt$_0:a - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - HashTable Sink Operator - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: b - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 3 (BROADCAST_EDGE) + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + input vertices: + 1 Map 3 + Statistics: Num rows: 791 Data size: 6328 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Map Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - Statistics: Num rows: 791 Data size: 6328 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: count() - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: bigint) - Execution mode: vectorized - Local Work: - Map Reduce Local Work - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join41.q.out ql/src/test/results/clientpositive/llap/join41.q.out new file mode 100644 index 0000000000..2bcbc21e88 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join41.q.out @@ -0,0 +1,208 @@ +PREHOOK: query: create table s1 as select * from src where key = 0 +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@s1 +POSTHOOK: query: create table s1 as select * from src where key = 0 +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@s1 +POSTHOOK: Lineage: s1.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: s1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: EXPLAIN +SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +PREHOOK: type: QUERY +PREHOOK: Input: default@s1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@s1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) > 10.0D) (type: boolean) + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) > 10.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +PREHOOK: type: QUERY +PREHOOK: Input: default@s1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@s1 +#### A masked pattern was here #### +0 val_0 NULL NULL +0 val_0 NULL NULL +0 val_0 NULL NULL +PREHOOK: query: EXPLAIN +SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +PREHOOK: type: QUERY +PREHOOK: Input: default@s1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@s1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (UDFToDouble(key) > 10.0D) (type: boolean) + Statistics: Num rows: 3 Data size: 522 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) > 10.0D) (type: boolean) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 3 Data size: 1044 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +PREHOOK: type: QUERY +PREHOOK: Input: default@s1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM s1 src1 LEFT OUTER JOIN s1 src2 ON (src1.key = src2.key AND src2.key > 10) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@s1 +#### A masked pattern was here #### +0 val_0 NULL NULL +0 val_0 NULL NULL +0 val_0 NULL NULL diff --git ql/src/test/results/clientpositive/llap/join42.q.out ql/src/test/results/clientpositive/llap/join42.q.out new file mode 100644 index 0000000000..743b828853 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join42.q.out @@ -0,0 +1,409 @@ +PREHOOK: query: create table L as select 4436 id +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@L +POSTHOOK: query: create table L as select 4436 id +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@L +POSTHOOK: Lineage: l.id SIMPLE [] +PREHOOK: query: create table LA_n11 as select 4436 loan_id, 4748 aid, 4415 pi_id +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@LA_n11 +POSTHOOK: query: create table LA_n11 as select 4436 loan_id, 4748 aid, 4415 pi_id +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@LA_n11 +POSTHOOK: Lineage: la_n11.aid SIMPLE [] +POSTHOOK: Lineage: la_n11.loan_id SIMPLE [] +POSTHOOK: Lineage: la_n11.pi_id SIMPLE [] +PREHOOK: query: create table FR as select 4436 loan_id +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@FR +POSTHOOK: query: create table FR as select 4436 loan_id +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@FR +POSTHOOK: Lineage: fr.loan_id SIMPLE [] +PREHOOK: query: create table A_n11 as select 4748 id +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@A_n11 +POSTHOOK: query: create table A_n11 as select 4748 id +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@A_n11 +POSTHOOK: Lineage: a_n11.id SIMPLE [] +PREHOOK: query: create table PI as select 4415 id +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@PI +POSTHOOK: query: create table PI as select 4415 id +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@PI +POSTHOOK: Lineage: pi.id SIMPLE [] +PREHOOK: query: create table acct as select 4748 aid, 10 acc_n, 122 brn +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: database:default +PREHOOK: Output: default@acct +POSTHOOK: query: create table acct as select 4748 aid, 10 acc_n, 122 brn +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: database:default +POSTHOOK: Output: default@acct +POSTHOOK: Lineage: acct.acc_n SIMPLE [] +POSTHOOK: Lineage: acct.aid SIMPLE [] +POSTHOOK: Lineage: acct.brn SIMPLE [] +PREHOOK: query: insert into table acct values(4748, null, null) +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@acct +POSTHOOK: query: insert into table acct values(4748, null, null) +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@acct +POSTHOOK: Lineage: acct.acc_n EXPRESSION [] +POSTHOOK: Lineage: acct.aid SCRIPT [] +POSTHOOK: Lineage: acct.brn EXPRESSION [] +PREHOOK: query: insert into table acct values(4748, null, null) +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@acct +POSTHOOK: query: insert into table acct values(4748, null, null) +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@acct +POSTHOOK: Lineage: acct.acc_n EXPRESSION [] +POSTHOOK: Lineage: acct.aid SCRIPT [] +POSTHOOK: Lineage: acct.brn EXPRESSION [] +Warning: Shuffle Join MERGEJOIN[86][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +Warning: Shuffle Join MERGEJOIN[87][tables = [$hdt$_4, $hdt$_5]] in Stage 'Reducer 10' is a cross product +PREHOOK: query: explain select + acct.ACC_N, + acct.brn +FROM L +JOIN LA_n11 ON L.id = LA_n11.loan_id +JOIN FR ON L.id = FR.loan_id +JOIN A_n11 ON LA_n11.aid = A_n11.id +JOIN PI ON PI.id = LA_n11.pi_id +JOIN acct ON A_n11.id = acct.aid +WHERE + L.id = 4436 + and acct.brn is not null +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n11 +PREHOOK: Input: default@acct +PREHOOK: Input: default@fr +PREHOOK: Input: default@l +PREHOOK: Input: default@la_n11 +PREHOOK: Input: default@pi +#### A masked pattern was here #### +POSTHOOK: query: explain select + acct.ACC_N, + acct.brn +FROM L +JOIN LA_n11 ON L.id = LA_n11.loan_id +JOIN FR ON L.id = FR.loan_id +JOIN A_n11 ON LA_n11.aid = A_n11.id +JOIN PI ON PI.id = LA_n11.pi_id +JOIN acct ON A_n11.id = acct.aid +WHERE + L.id = 4436 + and acct.brn is not null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n11 +POSTHOOK: Input: default@acct +POSTHOOK: Input: default@fr +POSTHOOK: Input: default@l +POSTHOOK: Input: default@la_n11 +POSTHOOK: Input: default@pi +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 10 <- Map 11 (XPROD_EDGE), Map 9 (XPROD_EDGE) + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) + Reducer 3 <- Map 7 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 8 (XPROD_EDGE), Reducer 3 (XPROD_EDGE) + Reducer 5 <- Reducer 10 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: la_n11 + filterExpr: ((loan_id = 4436) and aid is not null and pi_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((loan_id = 4436) and aid is not null and pi_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: aid (type: int), pi_id (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 11 + Map Operator Tree: + TableScan + alias: l + filterExpr: (id = 4436) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (id = 4436) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: a_n11 + filterExpr: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: id (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: pi + filterExpr: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: id (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 8 + Map Operator Tree: + TableScan + alias: fr + filterExpr: (loan_id = 4436) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (loan_id = 4436) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 9 + Map Operator Tree: + TableScan + alias: acct + filterExpr: (brn is not null and aid is not null) (type: boolean) + Statistics: Num rows: 3 Data size: 28 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (brn is not null and aid is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: aid (type: int), acc_n (type: int), brn (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 10 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int), _col2 (type: int) + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col2 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col2 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: int) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col6, _col7 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col6 (type: int), _col7 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[86][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +Warning: Shuffle Join MERGEJOIN[87][tables = [$hdt$_4, $hdt$_5]] in Stage 'Reducer 10' is a cross product +PREHOOK: query: select + acct.ACC_N, + acct.brn +FROM L +JOIN LA_n11 ON L.id = LA_n11.loan_id +JOIN FR ON L.id = FR.loan_id +JOIN A_n11 ON LA_n11.aid = A_n11.id +JOIN PI ON PI.id = LA_n11.pi_id +JOIN acct ON A_n11.id = acct.aid +WHERE + L.id = 4436 + and acct.brn is not null +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n11 +PREHOOK: Input: default@acct +PREHOOK: Input: default@fr +PREHOOK: Input: default@l +PREHOOK: Input: default@la_n11 +PREHOOK: Input: default@pi +#### A masked pattern was here #### +POSTHOOK: query: select + acct.ACC_N, + acct.brn +FROM L +JOIN LA_n11 ON L.id = LA_n11.loan_id +JOIN FR ON L.id = FR.loan_id +JOIN A_n11 ON LA_n11.aid = A_n11.id +JOIN PI ON PI.id = LA_n11.pi_id +JOIN acct ON A_n11.id = acct.aid +WHERE + L.id = 4436 + and acct.brn is not null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n11 +POSTHOOK: Input: default@acct +POSTHOOK: Input: default@fr +POSTHOOK: Input: default@l +POSTHOOK: Input: default@la_n11 +POSTHOOK: Input: default@pi +#### A masked pattern was here #### +10 122 diff --git ql/src/test/results/clientpositive/join43.q.out ql/src/test/results/clientpositive/llap/join43.q.out similarity index 50% rename from ql/src/test/results/clientpositive/join43.q.out rename to ql/src/test/results/clientpositive/llap/join43.q.out index 13696755f5..650754a584 100644 --- ql/src/test/results/clientpositive/join43.q.out +++ ql/src/test/results/clientpositive/llap/join43.q.out @@ -249,157 +249,154 @@ POSTHOOK: Input: default@events POSTHOOK: Input: default@purchase_history #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: cart_history - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), time (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: purchase_history - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), time (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Statistics: Num rows: 8 Data size: 744 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col3 > _col1) (type: boolean) - Statistics: Num rows: 2 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: max(_col1) - keys: _col2 (type: string), _col3 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: max(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: events - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), time (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: events + filterExpr: (s is not null and time is not null) (type: boolean) Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col2 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col2 (type: int) - Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col1 (type: int) - 1 _col0 (type: string), _col2 (type: int) - outputColumnNames: _col2 - Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), time (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: cart_history + filterExpr: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), time (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: purchase_history + filterExpr: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), time (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col1 (type: int) + 1 _col0 (type: string), _col2 (type: int) + outputColumnNames: _col2 + Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 85 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + residual filter predicates: {(_col3 > _col1)} + Statistics: Num rows: 2 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: max(_col1) + keys: _col2 (type: string), _col3 (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: max(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col2 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col2 (type: int) + Statistics: Num rows: 1 Data size: 89 Basic stats: COMPLETE Column stats: COMPLETE Stage: Stage-0 Fetch Operator @@ -489,155 +486,152 @@ POSTHOOK: Input: default@events POSTHOOK: Input: default@purchase_history #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-2 is a root stage - Stage-3 depends on stages: Stage-2 - Stage-1 depends on stages: Stage-3 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: cart_history - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), time (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - TableScan - alias: purchase_history - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), time (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col1, _col2, _col3 - Statistics: Num rows: 8 Data size: 744 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (_col3 > _col1) (type: boolean) - Statistics: Num rows: 2 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: max(_col1) - keys: _col2 (type: string), _col3 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: int) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: int) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: max(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: int) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: _col2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: events - filterExpr: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (s is not null and time is not null) (type: boolean) - Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: s (type: string), st2 (type: string), n (type: int), time (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col3 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col3 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: events + filterExpr: (s is not null and time is not null) (type: boolean) Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: int) - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col2 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col2 (type: int) - Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col3 (type: int) - 1 _col0 (type: string), _col2 (type: int) - outputColumnNames: _col1, _col2, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col4 (type: string), _col5 (type: int), _col6 (type: int), _col1 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), st2 (type: string), n (type: int), time (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col3 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col3 (type: int) + Statistics: Num rows: 6 Data size: 1086 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: cart_history + filterExpr: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), time (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 6 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: purchase_history + filterExpr: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (s is not null and time is not null) (type: boolean) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: s (type: string), time (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col3 (type: int) + 1 _col0 (type: string), _col2 (type: int) + outputColumnNames: _col1, _col2, _col4, _col5, _col6 + Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col4 (type: string), _col5 (type: int), _col6 (type: int), _col1 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 185 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2, _col3 + residual filter predicates: {(_col3 > _col1)} + Statistics: Num rows: 2 Data size: 186 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: max(_col1) + keys: _col2 (type: string), _col3 (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: int) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: int) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: max(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col2 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col2 (type: int) + Statistics: Num rows: 1 Data size: 93 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join44.q.out ql/src/test/results/clientpositive/llap/join44.q.out new file mode 100644 index 0000000000..e694995813 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join44.q.out @@ -0,0 +1,106 @@ +PREHOOK: query: CREATE TABLE mytable_n1(val1 INT, val2 INT, val3 INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@mytable_n1 +POSTHOOK: query: CREATE TABLE mytable_n1(val1 INT, val2 INT, val3 INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@mytable_n1 +PREHOOK: query: EXPLAIN +SELECT * +FROM mytable_n1 src1, mytable_n1 src2 +WHERE src1.val1=src2.val1 + AND src1.val2 between 2450816 and 2451500 + AND src2.val2 between 2450816 and 2451500 +PREHOOK: type: QUERY +PREHOOK: Input: default@mytable_n1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM mytable_n1 src1, mytable_n1 src2 +WHERE src1.val1=src2.val1 + AND src1.val2 between 2450816 and 2451500 + AND src2.val2 between 2450816 and 2451500 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@mytable_n1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: val1 (type: int) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + value expressions: val2 (type: int), val3 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (val1 is not null and val2 BETWEEN 2450816 AND 2451500) (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: val1 (type: int) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + value expressions: val2 (type: int), val3 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 val1 (type: int) + 1 val1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (_col0 = _col6) (type: boolean) + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join45.q.out ql/src/test/results/clientpositive/llap/join45.q.out new file mode 100644 index 0000000000..b35bfa75d4 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join45.q.out @@ -0,0 +1,2359 @@ +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 128 val_128 +128 128 val_128 +128 128 val_128 +146 val_146 146 val_146 +146 val_146 146 val_146 +150 val_150 150 val_150 +213 val_213 213 val_213 +213 val_213 213 val_213 +224 224 val_224 +224 224 val_224 +Warning: Shuffle Join MERGEJOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {((_col0 = _col3) or _col2 or _col5)} + Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 128 val_128 +128 128 val_128 +128 128 val_128 +146 val_146 146 val_146 +146 val_146 146 val_146 +150 val_150 150 val_150 +213 val_213 213 val_213 +213 val_213 213 val_213 +224 224 val_224 +224 224 val_224 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(_col2 + _col5) BETWEEN 100.0D AND 102.0D} + Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +66 val_66 34 val_34 +66 val_66 35 val_35 +66 val_66 35 val_35 +66 val_66 35 val_35 +98 val_98 2 val_2 +98 val_98 4 val_4 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(((_col2 + _col5) >= 100.0D) or ((_col2 + _col5) <= 102.0D))} + Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 0 val_0 +128 0 val_0 +128 0 val_0 +128 10 val_10 +128 100 val_100 +128 100 val_100 +128 103 val_103 +128 103 val_103 +128 104 val_104 +128 104 val_104 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(struct(_col2,_col5)) IN (const struct(100.0D,100.0D), const struct(101.0D,101.0D), const struct(102.0D,102.0D))} + Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 539 Data size: 102942 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 539 Data size: 102942 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {((UDFToDouble(_col4) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col4 (type: string), _col5 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col4 (type: string), _col5 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 213 val_213 NULL NULL +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +10 val_10 NULL NULL 66 val_66 +11 val_11 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 5 (XPROD_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {((UDFToDouble(_col2) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 354 Data size: 169615 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 val_128 128 128 +128 val_128 128 128 +128 val_128 128 128 +128 val_128 146 val_146 128 +128 val_128 146 val_146 128 +128 val_128 146 val_146 128 +128 val_128 150 val_150 128 +128 val_128 150 val_150 128 +128 val_128 150 val_150 128 +128 val_128 213 val_213 128 +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[18][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 5 (XPROD_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {((UDFToDouble(_col2) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4520 Data size: -2038608 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 528 Basic stats: PARTIAL Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 528 Basic stats: PARTIAL Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[18][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 213 val_213 NULL NULL +Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +10 val_10 NULL NULL 66 val_66 +11 val_11 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +Warning: Shuffle Join MERGEJOIN[37][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[38][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +Warning: Shuffle Join MERGEJOIN[40][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@cbo_t1 +PREHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@cbo_t1 +POSTHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 7 (XPROD_EDGE) + Reducer 3 <- Map 1 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (XPROD_EDGE), Reducer 6 (XPROD_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) + Reducer 6 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string), (c_int = 42) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean) + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: t3 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18 + Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30 + residual filter predicates: {((_col0 = _col25) or ((_col29 = _col17) and _col6))} {((_col13 = _col25) or ((_col28 = _col16) and (_col14 = _col26)))} {((_col7 = _col25) or ((_col8 = _col26) and (_col27 = _col21)))} + Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: +++++++++++++++++++++++++ + keys: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean) + null sort order: zzzzzzzzzzzzzzzzzzzzzzzzz + Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col30 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: float), _col16 (type: boolean), _col18 (type: string), _col19 (type: string), _col20 (type: int), _col21 (type: float), _col22 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: int), _col27 (type: float), _col28 (type: boolean) + null sort order: zzzzzzzzzzzzzzzzzzzzzzzzz + sort order: +++++++++++++++++++++++++ + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col5 (type: string), _col11 (type: string), _col17 (type: string), _col23 (type: string), _col29 (type: string) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: int), KEY.reducesinkkey3 (type: float), KEY.reducesinkkey4 (type: boolean), VALUE._col0 (type: string), KEY.reducesinkkey5 (type: string), KEY.reducesinkkey6 (type: string), KEY.reducesinkkey7 (type: int), KEY.reducesinkkey8 (type: float), KEY.reducesinkkey9 (type: boolean), VALUE._col1 (type: string), KEY.reducesinkkey10 (type: string), KEY.reducesinkkey11 (type: string), KEY.reducesinkkey12 (type: int), KEY.reducesinkkey13 (type: float), KEY.reducesinkkey14 (type: boolean), VALUE._col2 (type: string), KEY.reducesinkkey15 (type: string), KEY.reducesinkkey16 (type: string), KEY.reducesinkkey17 (type: int), KEY.reducesinkkey18 (type: float), KEY.reducesinkkey19 (type: boolean), VALUE._col3 (type: string), KEY.reducesinkkey20 (type: string), KEY.reducesinkkey21 (type: string), KEY.reducesinkkey22 (type: int), KEY.reducesinkkey23 (type: float), KEY.reducesinkkey24 (type: boolean), VALUE._col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col11 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[37][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[38][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +Warning: Shuffle Join MERGEJOIN[40][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +PREHOOK: query: SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@cbo_t1 +PREHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@cbo_t1 +POSTHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 diff --git ql/src/test/results/clientpositive/llap/join47.q.out ql/src/test/results/clientpositive/llap/join47.q.out new file mode 100644 index 0000000000..b35bfa75d4 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join47.q.out @@ -0,0 +1,2359 @@ +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(value) BETWEEN 100.0D AND 102.0D and key is not null) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 706 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + AND src1.value between 100 and 102 + AND src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 128 val_128 +128 128 val_128 +128 128 val_128 +146 val_146 146 val_146 +146 val_146 146 val_146 +150 val_150 150 val_150 +213 val_213 213 val_213 +213 val_213 213 val_213 +224 224 val_224 +224 224 val_224 +Warning: Shuffle Join MERGEJOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 350 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 38830 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[17][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.value between 100 and 102 + AND src.value between 100 and 102 + AND true) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4475 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(value) BETWEEN 100.0D AND 102.0D (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: boolean) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {((_col0 = _col3) or _col2 or _col5)} + Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 12500 Data size: 4512500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key=src.key + OR src1.value between 100 and 102 + OR src.value between 100 and 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 128 val_128 +128 128 val_128 +128 128 val_128 +146 val_146 146 val_146 +146 val_146 146 val_146 +150 val_150 150 val_150 +213 val_213 213 val_213 +213 val_213 213 val_213 +224 224 val_224 +224 224 val_224 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(_col2 + _col5) BETWEEN 100.0D AND 102.0D} + Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 1388 Data size: 512172 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1388 Data size: 489964 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + AND src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +66 val_66 34 val_34 +66 val_66 35 val_35 +66 val_66 35 val_35 +66 val_66 35 val_35 +98 val_98 2 val_2 +98 val_98 4 val_4 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(((_col2 + _col5) >= 100.0D) or ((_col2 + _col5) <= 102.0D))} + Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 8332 Data size: 3074508 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 8332 Data size: 2941196 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 3530 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON (src1.key+src.key >= 100 + OR src1.key+src.key <= 102) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 0 val_0 +128 0 val_0 +128 0 val_0 +128 10 val_10 +128 100 val_100 +128 100 val_100 +128 103 val_103 +128 103 val_103 +128 104 val_104 +128 104 val_104 +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 4 (XPROD_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {(struct(_col2,_col5)) IN (const struct(100.0D,100.0D), const struct(101.0D,101.0D), const struct(102.0D,102.0D))} + Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + null sort order: zzzz + Statistics: Num rows: 7 Data size: 2583 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzz + sort order: ++++ + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 7 Data size: 2471 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src1 JOIN src +ON ((src1.key,src.key) IN ((100,100),(101,101),(102,102))) +ORDER BY src1.key, src1.value, src.key, src.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 539 Data size: 102942 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 539 Data size: 102942 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + residual filter predicates: {((UDFToDouble(_col4) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col4 (type: string), _col5 (type: string), _col2 (type: string), _col3 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4491 Data size: 2342198 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col4 (type: string), _col5 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +LEFT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 213 val_213 NULL NULL +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 539 Data size: 107254 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4491 Data size: 2414054 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4491 Data size: 2312973 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +LEFT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +10 val_10 NULL NULL 66 val_66 +11 val_11 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 5 (XPROD_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {((UDFToDouble(_col2) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 354 Data size: 178440 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 354 Data size: 169615 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +RIGHT OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +128 val_128 128 128 +128 val_128 128 128 +128 val_128 128 128 +128 val_128 146 val_146 128 +128 val_128 146 val_146 128 +128 val_128 146 val_146 128 +128 val_128 150 val_150 128 +128 val_128 150 val_150 128 +128 val_128 150 val_150 128 +128 val_128 213 val_213 128 +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 64 Data size: 18640 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 533 Data size: 288650 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 533 Data size: 278754 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +RIGHT OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +Warning: Shuffle Join MERGEJOIN[18][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 5 (XPROD_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {((UDFToDouble(_col2) + UDFToDouble(_col0)) >= 100.0D)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4520 Data size: 174240 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4520 Data size: -2038608 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 528 Basic stats: PARTIAL Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 528 Basic stats: PARTIAL Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[18][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +JOIN src1 a ON (a.key+src.key >= 100) +FULL OUTER JOIN src1 b ON (b.key = src.key) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 128 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 146 val_146 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 150 val_150 NULL NULL +0 val_0 213 val_213 NULL NULL +Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4575 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 564 Data size: 14440 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: double), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + residual filter predicates: {((_col7 + _col2) <= 102.0D)} + Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++++++ + keys: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + null sort order: zzzzzz + Statistics: Num rows: 4700 Data size: 2493986 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string) + null sort order: zzzzzz + sort order: ++++++ + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: string), KEY.reducesinkkey3 (type: string), KEY.reducesinkkey4 (type: string), KEY.reducesinkkey5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 4700 Data size: 2358403 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 5280 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM src +FULL OUTER JOIN src1 a ON (a.key = src.key) +JOIN src1 b ON (b.key+src.key<= 102) +ORDER BY src.key, src.value, a.key, a.value, b.key, b.value +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 66 val_66 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +0 val_0 NULL NULL 98 val_98 +10 val_10 NULL NULL 66 val_66 +11 val_11 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +12 val_12 NULL NULL 66 val_66 +Warning: Shuffle Join MERGEJOIN[37][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[38][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +Warning: Shuffle Join MERGEJOIN[40][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +PREHOOK: query: EXPLAIN +SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@cbo_t1 +PREHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@cbo_t1 +POSTHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 7 (XPROD_EDGE) + Reducer 3 <- Map 1 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) + Reducer 4 <- Reducer 3 (XPROD_EDGE), Reducer 6 (XPROD_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) + Reducer 6 <- Map 1 (SIMPLE_EDGE), Map 7 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: t5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string), (c_int = 42) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7218 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean) + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 18 Data size: 6406 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: t3 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), c_int (type: int), c_float (type: float), c_boolean (type: boolean), dt (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 20 Data size: 7138 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 400 Data size: 294036 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18 + Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 8000 Data size: 8815454 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30 + residual filter predicates: {((_col0 = _col25) or ((_col29 = _col17) and _col6))} {((_col13 = _col25) or ((_col28 = _col16) and (_col14 = _col26)))} {((_col7 = _col25) or ((_col8 = _col26) and (_col27 = _col21)))} + Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: +++++++++++++++++++++++++ + keys: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean) + null sort order: zzzzzzzzzzzzzzzzzzzzzzzzz + Statistics: Num rows: 303750 Data size: 557077500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 10 + Select Operator + expressions: _col19 (type: string), _col20 (type: string), _col21 (type: int), _col22 (type: float), _col23 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: string), _col27 (type: int), _col28 (type: float), _col29 (type: boolean), _col30 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: int), _col10 (type: float), _col11 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: string), _col15 (type: int), _col16 (type: float), _col17 (type: boolean), _col18 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: float), _col16 (type: boolean), _col18 (type: string), _col19 (type: string), _col20 (type: int), _col21 (type: float), _col22 (type: boolean), _col24 (type: string), _col25 (type: string), _col26 (type: int), _col27 (type: float), _col28 (type: boolean) + null sort order: zzzzzzzzzzzzzzzzzzzzzzzzz + sort order: +++++++++++++++++++++++++ + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col5 (type: string), _col11 (type: string), _col17 (type: string), _col23 (type: string), _col29 (type: string) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: int), KEY.reducesinkkey3 (type: float), KEY.reducesinkkey4 (type: boolean), VALUE._col0 (type: string), KEY.reducesinkkey5 (type: string), KEY.reducesinkkey6 (type: string), KEY.reducesinkkey7 (type: int), KEY.reducesinkkey8 (type: float), KEY.reducesinkkey9 (type: boolean), VALUE._col1 (type: string), KEY.reducesinkkey10 (type: string), KEY.reducesinkkey11 (type: string), KEY.reducesinkkey12 (type: int), KEY.reducesinkkey13 (type: float), KEY.reducesinkkey14 (type: boolean), VALUE._col2 (type: string), KEY.reducesinkkey15 (type: string), KEY.reducesinkkey16 (type: string), KEY.reducesinkkey17 (type: int), KEY.reducesinkkey18 (type: float), KEY.reducesinkkey19 (type: boolean), VALUE._col3 (type: string), KEY.reducesinkkey20 (type: string), KEY.reducesinkkey21 (type: string), KEY.reducesinkkey22 (type: int), KEY.reducesinkkey23 (type: float), KEY.reducesinkkey24 (type: boolean), VALUE._col4 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29 + Statistics: Num rows: 303750 Data size: 555861675 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 10 + Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 10 Data size: 18300 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 90 Data size: 65601 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: int), _col3 (type: float), _col4 (type: boolean), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: int), _col9 (type: float), _col10 (type: boolean), _col11 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[37][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[38][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +Warning: Shuffle Join MERGEJOIN[40][tables = [$hdt$_0, $hdt$_1, $hdt$_2, $hdt$_3]] in Stage 'Reducer 4' is a cross product +PREHOOK: query: SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@cbo_t1 +PREHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### +POSTHOOK: query: SELECT * +FROM cbo_t1 t1 +RIGHT OUTER JOIN cbo_t1 t2 ON (t2.key = t1.key) +JOIN cbo_t1 t3 ON (t3.key = t2.key or t3.value = t2.value and t2.c_int = t1.c_int) +JOIN cbo_t1 t4 ON (t4.key = t2.key or t2.c_float = t4.c_float and t4.value = t2.value) +JOIN cbo_t1 t5 ON (t5.key = t2.key or t2.c_boolean = t4.c_boolean and t5.c_int = 42) +ORDER BY t1.key, t1.value, t1.c_int, t1.c_float, t1.c_boolean, t2.key, t2.value, t2.c_int, t2.c_float, t2.c_boolean, t3.key, t3.value, t3.c_int, t3.c_float, t3.c_boolean, t4.key, t4.value, t4.c_int, t4.c_float, t4.c_boolean, t5.key, t5.value, t5.c_int, t5.c_float, t5.c_boolean +LIMIT 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@cbo_t1 +POSTHOOK: Input: default@cbo_t1@dt=2014 +#### A masked pattern was here #### + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 + 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 1 1 1 1.0 true 2014 diff --git ql/src/test/results/clientpositive/llap/join5.q.out ql/src/test/results/clientpositive/llap/join5.q.out new file mode 100644 index 0000000000..dce3d2eb06 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join5.q.out @@ -0,0 +1,229 @@ +PREHOOK: query: CREATE TABLE dest1_n126(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n126 +POSTHOOK: query: CREATE TABLE dest1_n126(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n126 +PREHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + RIGHT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n126 +POSTHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + RIGHT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n126 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n126 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) + outputColumnNames: c1, c2, c3, c4 + Statistics: Num rows: 55 Data size: 10450 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') + minReductionHashAggr: 0.9818182 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n126 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3, c4 + Column Types: int, string, int, string + Table: default.dest1_n126 + +PREHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + RIGHT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n126 +POSTHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + RIGHT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n126 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n126 +POSTHOOK: Lineage: dest1_n126.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n126.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n126.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n126.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n126.* FROM dest1_n126 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n126 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n126.* FROM dest1_n126 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n126 +#### A masked pattern was here #### +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +NULL NULL 20 val_20 +NULL NULL 24 val_24 +NULL NULL 24 val_24 diff --git ql/src/test/results/clientpositive/llap/join6.q.out ql/src/test/results/clientpositive/llap/join6.q.out new file mode 100644 index 0000000000..e20d166590 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join6.q.out @@ -0,0 +1,234 @@ +PREHOOK: query: CREATE TABLE dest1_n156(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n156 +POSTHOOK: query: CREATE TABLE dest1_n156(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n156 +PREHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n156 +POSTHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n156 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n156 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) + outputColumnNames: c1, c2, c3, c4 + Statistics: Num rows: 110 Data size: 20900 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n156 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3, c4 + Column Types: int, string, int, string + Table: default.dest1_n156 + +PREHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n156 +POSTHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n156 SELECT c.c1, c.c2, c.c3, c.c4 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n156 +POSTHOOK: Lineage: dest1_n156.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n156.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n156.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n156.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n156.* FROM dest1_n156 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n156 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n156.* FROM dest1_n156 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n156 +#### A masked pattern was here #### +11 val_11 NULL NULL +12 val_12 NULL NULL +12 val_12 NULL NULL +15 val_15 NULL NULL +15 val_15 NULL NULL +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +NULL NULL 20 val_20 +NULL NULL 24 val_24 +NULL NULL 24 val_24 diff --git ql/src/test/results/clientpositive/llap/join7.q.out ql/src/test/results/clientpositive/llap/join7.q.out new file mode 100644 index 0000000000..2f4c862200 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join7.q.out @@ -0,0 +1,289 @@ +PREHOOK: query: CREATE TABLE dest1_n17(c1 INT, c2 STRING, c3 INT, c4 STRING, c5 INT, c6 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n17 +POSTHOOK: query: CREATE TABLE dest1_n17(c1 INT, c2 STRING, c3 INT, c4 STRING, c5 INT, c6 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n17 +PREHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + LEFT OUTER JOIN + ( + FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 + ) c + ON (a.c1 = c.c5) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 +) c +INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n17 +POSTHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + LEFT OUTER JOIN + ( + FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 + ) c + ON (a.c1 = c.c5) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 +) c +INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n17 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) or ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 25.0D))) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Filter Operator + predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 25.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 110 Data size: 39160 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 110 Data size: 58740 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), UDFToInteger(_col2) (type: int), _col3 (type: string), UDFToInteger(_col4) (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n17 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string), _col4 (type: int), _col5 (type: string) + outputColumnNames: c1, c2, c3, c4, c5, c6 + Statistics: Num rows: 110 Data size: 31350 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll'), compute_stats(c5, 'hll'), compute_stats(c6, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 2592 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 2592 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4), compute_stats(VALUE._col5) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 2640 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 2640 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n17 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3, c4, c5, c6 + Column Types: int, string, int, string, int, string + Table: default.dest1_n17 + +PREHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + LEFT OUTER JOIN + ( + FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 + ) c + ON (a.c1 = c.c5) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 +) c +INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n17 +POSTHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + FULL OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + LEFT OUTER JOIN + ( + FROM src src3 SELECT src3.key AS c5, src3.value AS c6 WHERE src3.key > 20 and src3.key < 25 + ) c + ON (a.c1 = c.c5) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4, c.c5 AS c5, c.c6 AS c6 +) c +INSERT OVERWRITE TABLE dest1_n17 SELECT c.c1, c.c2, c.c3, c.c4, c.c5, c.c6 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n17 +POSTHOOK: Lineage: dest1_n17.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n17.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n17.c3 EXPRESSION [(src)src2.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n17.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n17.c5 EXPRESSION [(src)src3.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n17.c6 SIMPLE [(src)src3.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n17.* FROM dest1_n17 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n17 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n17.* FROM dest1_n17 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n17 +#### A masked pattern was here #### +11 val_11 NULL NULL NULL NULL +12 val_12 NULL NULL NULL NULL +12 val_12 NULL NULL NULL NULL +15 val_15 NULL NULL NULL NULL +15 val_15 NULL NULL NULL NULL +17 val_17 17 val_17 NULL NULL +18 val_18 18 val_18 NULL NULL +18 val_18 18 val_18 NULL NULL +18 val_18 18 val_18 NULL NULL +18 val_18 18 val_18 NULL NULL +19 val_19 19 val_19 NULL NULL +NULL NULL 20 val_20 NULL NULL +NULL NULL 24 val_24 NULL NULL +NULL NULL 24 val_24 NULL NULL diff --git ql/src/test/results/clientpositive/llap/join8.q.out ql/src/test/results/clientpositive/llap/join8.q.out new file mode 100644 index 0000000000..a7f500baa3 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join8.q.out @@ -0,0 +1,228 @@ +PREHOOK: query: CREATE TABLE dest1_n173(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest1_n173 +POSTHOOK: query: CREATE TABLE dest1_n173(c1 INT, c2 STRING, c3 INT, c4 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest1_n173 +PREHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n173 +POSTHOOK: query: EXPLAIN +FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n173 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 10.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col2 is null (type: boolean) + Statistics: Num rows: 1 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string), null (type: int), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n173 + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col2 (type: int), _col3 (type: string) + outputColumnNames: c1, c2, c3, c4 + Statistics: Num rows: 1 Data size: 190 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(c1, 'hll'), compute_stats(c2, 'hll'), compute_stats(c3, 'hll'), compute_stats(c4, 'hll') + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1728 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n173 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: c1, c2, c3, c4 + Column Types: int, string, int, string + Table: default.dest1_n173 + +PREHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest1_n173 +POSTHOOK: query: FROM ( + FROM + ( + FROM src src1 SELECT src1.key AS c1, src1.value AS c2 WHERE src1.key > 10 and src1.key < 20 + ) a + LEFT OUTER JOIN + ( + FROM src src2 SELECT src2.key AS c3, src2.value AS c4 WHERE src2.key > 15 and src2.key < 25 + ) b + ON (a.c1 = b.c3) + SELECT a.c1 AS c1, a.c2 AS c2, b.c3 AS c3, b.c4 AS c4 +) c +INSERT OVERWRITE TABLE dest1_n173 SELECT c.c1, c.c2, c.c3, c.c4 where c.c3 IS NULL AND c.c1 IS NOT NULL +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest1_n173 +POSTHOOK: Lineage: dest1_n173.c1 EXPRESSION [(src)src1.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n173.c2 SIMPLE [(src)src1.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: dest1_n173.c3 EXPRESSION [] +POSTHOOK: Lineage: dest1_n173.c4 SIMPLE [(src)src2.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: SELECT dest1_n173.* FROM dest1_n173 +PREHOOK: type: QUERY +PREHOOK: Input: default@dest1_n173 +#### A masked pattern was here #### +POSTHOOK: query: SELECT dest1_n173.* FROM dest1_n173 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dest1_n173 +#### A masked pattern was here #### +11 val_11 NULL NULL +12 val_12 NULL NULL +12 val_12 NULL NULL +15 val_15 NULL NULL +15 val_15 NULL NULL diff --git ql/src/test/results/clientpositive/join9.q.out ql/src/test/results/clientpositive/llap/join9.q.out similarity index 59% rename from ql/src/test/results/clientpositive/join9.q.out rename to ql/src/test/results/clientpositive/llap/join9.q.out index 0479d649a0..d28d5fe047 100644 --- ql/src/test/results/clientpositive/join9.q.out +++ ql/src/test/results/clientpositive/llap/join9.q.out @@ -31,187 +31,134 @@ FROM `default`.`src` WHERE `key` IS NOT NULL) AS `t2` ON `t0`.`key` = `t2`.`key` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src1 - filterExpr: ((ds = '2008-04-08') and (hr = '12') and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - auto parallelism: false - TableScan - alias: src2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Tez #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src + Path -> Partition: #### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:src2] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:src1] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=12 [src1] + Map 4 + Map Operator Tree: + TableScan + alias: src2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + Path -> Partition: #### A masked pattern was here #### - table: + Partition + base file name: src input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -220,53 +167,152 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments - columns.types int:string + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.dest1_n39 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n39 { i32 key, string value} + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 + totalSize 5812 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n39 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: int), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src2] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Position of Big Table: 1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.dest1_n39 + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct dest1_n39 { i32 key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n39 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 791 Data size: 75145 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator - bucketingVersion: 1 + bucketingVersion: 2 compressed: false GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - column.name.delimiter , + bucketing_version -1 columns _col0,_col1 - columns.types struct,struct + columns.types struct:struct escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe TotalFiles: 1 GatherStats: false MultiFileSpray: false + Stage: Stage-2 + Dependency Collection + Stage: Stage-0 Move Operator tables: @@ -296,7 +342,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n39 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -306,80 +352,6 @@ STAGE PLANS: Table: default.dest1_n39 Is Table Level Stats: true - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10002 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1 - columns.types struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - PREHOOK: query: FROM srcpart src1 JOIN src src2 ON (src1.key = src2.key) INSERT OVERWRITE TABLE dest1_n39 SELECT src1.key, src2.value where src1.ds = '2008-04-08' and src1.hr = '12' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/join_1to1.q.out ql/src/test/results/clientpositive/llap/join_1to1.q.out similarity index 96% rename from ql/src/test/results/clientpositive/join_1to1.q.out rename to ql/src/test/results/clientpositive/llap/join_1to1.q.out index 1ccb613763..5619687d7d 100644 --- ql/src/test/results/clientpositive/join_1to1.q.out +++ ql/src/test/results/clientpositive/llap/join_1to1.q.out @@ -132,6 +132,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -142,22 +143,17 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 @@ -222,6 +218,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -232,22 +229,17 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 @@ -352,6 +344,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -362,22 +355,18 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 @@ -442,6 +431,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -452,22 +442,18 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 @@ -572,6 +558,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -583,17 +570,22 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 +NULL NULL NULL 40 10040 88 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 +NULL NULL NULL 70 10040 88 +NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 80 10040 66 @@ -662,6 +654,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -673,17 +666,22 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 +NULL NULL NULL 40 10040 88 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 +NULL NULL NULL 70 10040 88 +NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 80 10040 66 @@ -792,6 +790,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -802,22 +801,17 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 @@ -882,6 +876,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -892,22 +887,17 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 @@ -1012,6 +1002,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -1022,22 +1013,18 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a full outer join join_1to1_2 b on a.key1 = b.key1 and a.key2 = b.key2 @@ -1102,6 +1089,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -1112,22 +1100,18 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 60 10040 66 -70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 -NULL NULL NULL 80 10040 66 -NULL NULL NULL 80 10040 66 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 PREHOOK: query: SELECT * FROM join_1to1_1 a join join_1to1_2 b on a.key1 = b.key1 @@ -1232,6 +1216,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -1243,17 +1228,22 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 +NULL NULL NULL 40 10040 88 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 +NULL NULL NULL 70 10040 88 +NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 80 10040 66 @@ -1322,6 +1312,7 @@ POSTHOOK: Input: default@join_1to1_2 30 10030 66 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 66 40 10040 66 +40 10040 66 NULL NULL NULL 40 10040 88 NULL NULL NULL 5 10005 66 5 10005 66 50 10050 66 50 10050 66 @@ -1333,17 +1324,22 @@ POSTHOOK: Input: default@join_1to1_2 60 10040 66 60 10040 66 60 10040 66 60 10040 66 70 10040 66 NULL NULL NULL -70 10040 66 NULL NULL NULL 80 10040 88 NULL NULL NULL 80 10040 88 NULL NULL NULL NULL 10050 66 NULL NULL NULL NULL NULL 66 NULL NULL NULL NULL NULL NULL 10 10010 66 -NULL NULL NULL 25 10025 66 NULL NULL NULL 30 10030 88 NULL NULL NULL 35 10035 88 NULL NULL NULL 40 10040 88 +NULL NULL NULL 40 10040 88 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 66 +NULL NULL NULL 50 10050 88 NULL NULL NULL 50 10050 88 +NULL NULL NULL 50 10050 88 +NULL NULL NULL 70 10040 88 +NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 70 10040 88 NULL NULL NULL 80 10040 66 diff --git ql/src/test/results/clientpositive/llap/join_alt_syntax.q.out ql/src/test/results/clientpositive/llap/join_alt_syntax.q.out new file mode 100644 index 0000000000..ad7b562a62 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_alt_syntax.q.out @@ -0,0 +1,765 @@ +Warning: Shuffle Join MERGEJOIN[9][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: explain select p1.p_name, p2.p_name +from part p1 , part p2 +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name +from part p1 , part p2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: p2 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1 + Statistics: Num rows: 676 Data size: 163592 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 676 Data size: 163592 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 ,part p2 ,part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 ,part p2 ,part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 , (select p_name from part) p2 ,part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 , (select p_name from part) p2 ,part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: part + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 6534 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 10164 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_0, $hdt$_1, $hdt$_2]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 , part p2 , part p3 +where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name +from part p1 , part p2 , part p3 +where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + residual filter predicates: {((_col0 + _col3) = _col3)} + Statistics: Num rows: 351 Data size: 130221 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col4 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 351 Data size: 127413 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 351 Data size: 127413 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name +from part p1 , part p2 join part p3 on p2.p_name = p1.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name +from part p1 , part p2 join part p3 on p2.p_name = p1.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p4 + filterExpr: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col1, _col3, _col5, _col6 + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name +from part p1 join part p2 on p2.p_name = p1.p_name , part p3 , part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_name, p2.p_name, p3.p_name, p4.p_name +from part p1 join part p2 on p2.p_name = p1.p_name , part p3 , part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p4 + filterExpr: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 3250 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 26 Data size: 3146 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col1, _col3, _col5, _col6 + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 14036 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 6642 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_by_range_rule_not_null.q.out ql/src/test/results/clientpositive/llap/join_by_range_rule_not_null.q.out new file mode 100644 index 0000000000..8d3e049912 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_by_range_rule_not_null.q.out @@ -0,0 +1,518 @@ +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 39 Data size: 13767 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[15][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key < b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key < b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {(_col0 < _col2)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key AND a.value >= b.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key = b.key AND a.value >= b.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {(_col1 >= _col3)} + Statistics: Num rows: 13 Data size: 4589 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 13 Data size: 4589 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[15][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + residual filter predicates: {(_col0 > _col3)} + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 4166 Data size: 1470598 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[9][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.key OR 1 = 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key > b.key OR 1 = 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 12500 Data size: 4412500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[7][tables = [a, b]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key IS DISTINCT FROM b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM src a JOIN src1 b ON a.key IS DISTINCT FROM b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 25 Data size: 4375 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col5, _col6 + residual filter predicates: {(not (_col0 IS NOT DISTINCT FROM _col5))} + Statistics: Num rows: 6250 Data size: 2206250 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 6250 Data size: 2206250 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6250 Data size: 2206250 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_1.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_1.q.out new file mode 100644 index 0000000000..d415c6245d --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_1.q.out @@ -0,0 +1,477 @@ +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name and p2.p_name = p3.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name and p2.p_name = p3.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + residual filter predicates: {((_col0 + _col18) = _col18)} + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_partkey = 1 and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_partkey = 1 and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p_partkey = 1) and p_name is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((p_partkey = 1) and p_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 619 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 + Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 + Statistics: Num rows: 26 Data size: 48178 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_2.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_2.q.out new file mode 100644 index 0000000000..e0d62a5007 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_2.q.out @@ -0,0 +1,310 @@ +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 on p2.p_name = p3.p_name and p1.p_name = p4.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 on p2.p_name = p3.p_name and p1.p_name = p4.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 on p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 on p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: (p_name is not null or (p_name is not null and p_partkey is not null)) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p_name is not null and p_partkey is not null) or p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col9 (type: int), _col10 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col9 (type: int), _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col9 (type: int), _col10 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_3.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_3.q.out new file mode 100644 index 0000000000..78b20d11d1 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_3.q.out @@ -0,0 +1,485 @@ +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p1.p_name = p2.p_name and p2.p_name = p3.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_name = p1.p_name and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_name = p1.p_name and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 51996 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_partkey + p1.p_partkey = p1.p_partkey and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + residual filter predicates: {((_col0 + _col18) = _col18)} + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 351 Data size: 651807 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_partkey = 1 and p3.p_name = p2.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 +where p2.p_partkey = 1 and p3.p_name = p2.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 4 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p_partkey = 1) and p_name is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((p_partkey = 1) and p_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 619 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 615 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 + Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1234 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 + Statistics: Num rows: 26 Data size: 48178 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 26 Data size: 48282 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_4.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_4.q.out new file mode 100644 index 0000000000..70373d5f71 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_4.q.out @@ -0,0 +1,314 @@ +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_name = p4.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p1.p_name = p2.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_name = p4.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part p2 join part p3 on p2.p_name = p1.p_name join part p4 +where p2.p_name = p3.p_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2.p_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: (p_name is not null or (p_name is not null and p_partkey is not null)) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p_name is not null and p_partkey is not null) or p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col9 (type: int), _col10 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col9 (type: int), _col10 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col9 (type: int), _col10 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 71804 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 27 Data size: 33426 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual1.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual1.q.out new file mode 100644 index 0000000000..bf48209735 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual1.q.out @@ -0,0 +1,603 @@ +PREHOOK: query: create table part2_n0( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part2_n0 +POSTHOOK: query: create table part2_n0( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part2_n0 +PREHOOK: query: create table part3_n0( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part3_n0 +POSTHOOK: query: create table part3_n0( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part3_n0 +PREHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p1.p_name = p2_name and p2_name = p3_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n0 +PREHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p1.p_name = p2_name and p2_name = p3_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n0 +POSTHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_name = p1.p_name and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n0 +PREHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_name = p1.p_name and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n0 +POSTHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey + p_partkey = p1.p_partkey and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n0 +PREHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey + p_partkey = p1.p_partkey and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n0 +POSTHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + residual filter predicates: {((_col0 + _col18) = _col18)} + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey = 1 and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n0 +PREHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n0 p2 join part3_n0 p3 on p2_partkey = 1 and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n0 +POSTHOOK: Input: default@part3_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p2_partkey = 1) and p2_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((p2_partkey = 1) and p2_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual2.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual2.q.out new file mode 100644 index 0000000000..c570eca4cb --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual2.q.out @@ -0,0 +1,397 @@ +PREHOOK: query: create table part2( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part2 +POSTHOOK: query: create table part2( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part2 +PREHOOK: query: create table part3( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part3 +POSTHOOK: query: create table part3( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part3 +PREHOOK: query: explain select * +from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 on p2_name = p3_name and p1.p_name = p4.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2 +PREHOOK: Input: default@part3 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2 p2 join part3 p3 on p1.p_name = p2_name join part p4 on p2_name = p3_name and p1.p_name = p4.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2 +POSTHOOK: Input: default@part3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 6 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col19 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col19 (type: string) + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col18 (type: int), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col19 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 on p2_name = p3_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2 +PREHOOK: Input: default@part3 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2 p2 join part3 p3 on p2_name = p1.p_name join part p4 on p2_name = p3_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2 +POSTHOOK: Input: default@part3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 6 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: (p2_name is not null and p2_partkey is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (p2_name is not null and p2_partkey is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: ((p_name is not null and p_partkey is not null) or p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col18 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col18 (type: int) + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col18 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual3.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual3.q.out new file mode 100644 index 0000000000..00c8519c2e --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual3.q.out @@ -0,0 +1,611 @@ +PREHOOK: query: create table part2_n5( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part2_n5 +POSTHOOK: query: create table part2_n5( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part2_n5 +PREHOOK: query: create table part3_n2( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part3_n2 +POSTHOOK: query: create table part3_n2( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part3_n2 +PREHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p1.p_name = p2_name and p2_name = p3_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n5 +PREHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p1.p_name = p2_name and p2_name = p3_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n5 +POSTHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_name = p1.p_name and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n5 +PREHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_name = p1.p_name and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n5 +POSTHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n5 +PREHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_partkey + p1.p_partkey = p1.p_partkey and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n5 +POSTHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + residual filter predicates: {((_col0 + _col18) = _col18)} + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 13 Data size: 24076 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +Warning: Shuffle Join MERGEJOIN[31][tables = [$hdt$_1, $hdt$_2, $hdt$_0]] in Stage 'Reducer 3' is a cross product +PREHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_partkey = 1 and p3_name = p2_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n5 +PREHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n5 p2 join part3_n2 p3 +where p2_partkey = 1 and p3_name = p2_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n5 +POSTHOOK: Input: default@part3_n2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (XPROD_EDGE), Reducer 2 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: ((p2_partkey = 1) and p2_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((p2_partkey = 1) and p2_name is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: int), _col5 (type: string), _col6 (type: double), _col7 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25 + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col17 (type: int), _col18 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: int), _col23 (type: string), _col24 (type: double), _col25 (type: string), 1 (type: int), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: int), _col14 (type: string), _col15 (type: double), _col16 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 26 Data size: 48152 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual4.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual4.q.out new file mode 100644 index 0000000000..ee032d9e5e --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual4.q.out @@ -0,0 +1,401 @@ +PREHOOK: query: create table part2_n4( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part2_n4 +POSTHOOK: query: create table part2_n4( + p2_partkey INT, + p2_name STRING, + p2_mfgr STRING, + p2_brand STRING, + p2_type STRING, + p2_size INT, + p2_container STRING, + p2_retailprice DOUBLE, + p2_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part2_n4 +PREHOOK: query: create table part3_n1( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@part3_n1 +POSTHOOK: query: create table part3_n1( + p3_partkey INT, + p3_name STRING, + p3_mfgr STRING, + p3_brand STRING, + p3_type STRING, + p3_size INT, + p3_container STRING, + p3_retailprice DOUBLE, + p3_comment STRING +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@part3_n1 +PREHOOK: query: explain select * +from part p1 join part2_n4 p2 join part3_n1 p3 on p1.p_name = p2_name join part p4 +where p2_name = p3_name and p1.p_name = p4.p_name +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n4 +PREHOOK: Input: default@part3_n1 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n4 p2 join part3_n1 p3 on p1.p_name = p2_name join part p4 +where p2_name = p3_name and p1.p_name = p4.p_name +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n4 +POSTHOOK: Input: default@part3_n1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 6 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p2_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_name is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col19 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col19 (type: string) + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col18 (type: int), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col19 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * +from part p1 join part2_n4 p2 join part3_n1 p3 on p2_name = p1.p_name join part p4 +where p2_name = p3_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2_partkey +PREHOOK: type: QUERY +PREHOOK: Input: default@part +PREHOOK: Input: default@part2_n4 +PREHOOK: Input: default@part3_n1 +#### A masked pattern was here #### +POSTHOOK: query: explain select * +from part p1 join part2_n4 p2 join part3_n1 p3 on p2_name = p1.p_name join part p4 +where p2_name = p3_name and p1.p_partkey = p4.p_partkey + and p1.p_partkey = p2_partkey +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +POSTHOOK: Input: default@part2_n4 +POSTHOOK: Input: default@part3_n1 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 6 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: (p2_name is not null and p2_partkey is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (p2_name is not null and p2_partkey is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p2_partkey (type: int), p2_name (type: string), p2_mfgr (type: string), p2_brand (type: string), p2_type (type: string), p2_size (type: int), p2_container (type: string), p2_retailprice (type: double), p2_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p3 + filterExpr: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: p3_name is not null (type: boolean) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: p3_partkey (type: int), p3_name (type: string), p3_mfgr (type: string), p3_brand (type: string), p3_type (type: string), p3_size (type: int), p3_container (type: string), p3_retailprice (type: double), p3_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 1120 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: ((p_name is not null and p_partkey is not null) or p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_name is not null and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_name (type: string), p_mfgr (type: string), p_brand (type: string), p_type (type: string), p_size (type: int), p_container (type: string), p_retailprice (type: double), p_comment (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 16094 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17 + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1232 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: string) + 1 _col0 (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26 + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col18 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col18 (type: int) + Statistics: Num rows: 28 Data size: 17703 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col18 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col18 (type: int), _col19 (type: string), _col20 (type: string), _col21 (type: string), _col22 (type: string), _col23 (type: int), _col24 (type: string), _col25 (type: double), _col26 (type: string), _col0 (type: int), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: int), _col6 (type: string), _col7 (type: double), _col8 (type: string), _col9 (type: int), _col10 (type: string), _col11 (type: string), _col12 (type: string), _col13 (type: string), _col14 (type: int), _col15 (type: string), _col16 (type: double), _col17 (type: string), _col27 (type: int), _col28 (type: string), _col29 (type: string), _col30 (type: string), _col31 (type: string), _col32 (type: int), _col33 (type: string), _col34 (type: double), _col35 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13, _col14, _col15, _col16, _col17, _col18, _col19, _col20, _col21, _col22, _col23, _col24, _col25, _col26, _col27, _col28, _col29, _col30, _col31, _col32, _col33, _col34, _col35 + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 30 Data size: 19473 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual5.q.out ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual5.q.out new file mode 100644 index 0000000000..8183f2cff5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_cond_pushdown_unqual5.q.out @@ -0,0 +1,233 @@ +PREHOOK: query: CREATE TABLE ltable (index int, la int, lk1 string, lk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@ltable +POSTHOOK: query: CREATE TABLE ltable (index int, la int, lk1 string, lk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@ltable +PREHOOK: query: CREATE TABLE rtable (ra int, rk1 string, rk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@rtable +POSTHOOK: query: CREATE TABLE rtable (ra int, rk1 string, rk2 string) ROW FORMAT DELIMITED FIELDS TERMINATED BY ',' +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@rtable +PREHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@ltable +POSTHOOK: query: insert into ltable values (1, null, 'CD5415192314304', '00071'), (2, null, 'CD5415192225530', '00071') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@ltable +POSTHOOK: Lineage: ltable.index SCRIPT [] +POSTHOOK: Lineage: ltable.la EXPRESSION [] +POSTHOOK: Lineage: ltable.lk1 SCRIPT [] +POSTHOOK: Lineage: ltable.lk2 SCRIPT [] +PREHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@rtable +POSTHOOK: query: insert into rtable values (1, 'CD5415192314304', '00071'), (45, 'CD5415192314304', '00072') +POSTHOOK: type: QUERY +POSTHOOK: Input: _dummy_database@_dummy_table +POSTHOOK: Output: default@rtable +POSTHOOK: Lineage: rtable.ra SCRIPT [] +POSTHOOK: Lineage: rtable.rk1 SCRIPT [] +POSTHOOK: Lineage: rtable.rk2 SCRIPT [] +PREHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +PREHOOK: type: QUERY +PREHOOK: Input: default@ltable +PREHOOK: Input: default@rtable +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ltable +POSTHOOK: Input: default@rtable +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: l + Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: index (type: int), la (type: int), lk1 (type: string), lk2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) + Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: r + filterExpr: (rk1 is not null and rk2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (rk1 is not null and rk2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ra (type: int), rk1 (type: string), rk2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string), _col2 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col1 (type: string), _col2 (type: string) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col2 (type: string), _col3 (type: string) + 1 _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 3 Data size: 1160 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: CASE WHEN (_col1 is not null) THEN (CASE WHEN (_col4 is not null) THEN ((CAST( _col4 AS STRING) = CAST( _col1 AS STRING))) ELSE (('EMPTY' = CAST( _col1 AS STRING))) END) ELSE (CASE WHEN (_col4 is not null) THEN ((CAST( _col4 AS STRING) = 'EMPTY')) ELSE (('EMPTY' = 'EMPTY')) END) END (type: boolean) + Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +PREHOOK: type: QUERY +PREHOOK: Input: default@ltable +PREHOOK: Input: default@rtable +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ltable +POSTHOOK: Input: default@rtable +#### A masked pattern was here #### +2 NULL CD5415192225530 00071 NULL NULL NULL +PREHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +PREHOOK: type: QUERY +PREHOOK: Input: default@ltable +PREHOOK: Input: default@rtable +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ltable +POSTHOOK: Input: default@rtable +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: l + Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: index (type: int), la (type: int), lk1 (type: string), lk2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col2 (type: string), _col3 (type: string) + 1 _col1 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + input vertices: + 1 Map 2 + Statistics: Num rows: 3 Data size: 1160 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: CASE WHEN (_col1 is not null) THEN (CASE WHEN (_col4 is not null) THEN ((CAST( _col4 AS STRING) = CAST( _col1 AS STRING))) ELSE (('EMPTY' = CAST( _col1 AS STRING))) END) ELSE (CASE WHEN (_col4 is not null) THEN ((CAST( _col4 AS STRING) = 'EMPTY')) ELSE (('EMPTY' = 'EMPTY')) END) END (type: boolean) + Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 388 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: r + filterExpr: (rk1 is not null and rk2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (rk1 is not null and rk2 is not null) (type: boolean) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: ra (type: int), rk1 (type: string), rk2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string), _col2 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col1 (type: string), _col2 (type: string) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +PREHOOK: type: QUERY +PREHOOK: Input: default@ltable +PREHOOK: Input: default@rtable +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM ltable l LEFT OUTER JOIN rtable r on (l.lk1 = r.rk1 AND l.lk2 = r.rk2) WHERE COALESCE(l.la,'EMPTY')=COALESCE(r.ra,'EMPTY') +POSTHOOK: type: QUERY +POSTHOOK: Input: default@ltable +POSTHOOK: Input: default@rtable +#### A masked pattern was here #### +2 NULL CD5415192225530 00071 NULL NULL NULL diff --git ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out new file mode 100644 index 0000000000..dd0f0337e9 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out @@ -0,0 +1,1652 @@ +PREHOOK: query: create table a_n4 as SELECT 100 as key, a_n4.value as value FROM src LATERAL VIEW explode(array(40, 50, 60)) a_n4 as value limit 3 +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@a_n4 +POSTHOOK: query: create table a_n4 as SELECT 100 as key, a_n4.value as value FROM src LATERAL VIEW explode(array(40, 50, 60)) a_n4 as value limit 3 +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@a_n4 +POSTHOOK: Lineage: a_n4.key SIMPLE [] +POSTHOOK: Lineage: a_n4.value SCRIPT [] +PREHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` +FROM (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` +FROM `default`.`a_n4`) AS `t` +LEFT JOIN (SELECT `key`, CAST(50 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 50 AND `key` IS NOT NULL) AS `t1` ON `t`.`key` = `t1`.`key` AND `t`.`=3` +LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t`.`key` = `t3`.`key` AND `t`.`=` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n4 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 60) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 60 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [a_n4] + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 50 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col2} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col4, _col5 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col4 (type: int), _col5 (type: int) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col1} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col4, _col5, _col6, _col7 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5 + columns.types int:int:int:int:int:int + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 40 NULL NULL NULL NULL +100 50 100 50 NULL NULL +100 60 NULL NULL 100 60 +PREHOOK: query: select /*+ MAPJOIN(b,c)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ MAPJOIN(b,c)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 40 NULL NULL NULL NULL +100 50 100 50 NULL NULL +100 60 NULL NULL 100 60 +PREHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` +FROM (SELECT `key`, CAST(50 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 50 AND `key` IS NOT NULL) AS `t0` +RIGHT JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` +FROM `default`.`a_n4`) AS `t1` ON `t0`.`key` = `t1`.`key` AND `t1`.`=3` +LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t1`.`key` = `t3`.`key` AND `t1`.`=` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n4 + filterExpr: (((value = 50) and key is not null) or ((value = 60) and key is not null)) (type: boolean) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 50 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 60) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 60 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [a_n4] + Map 4 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter mappings: + 1 [0, 1] + filter predicates: + 0 + 1 {VALUE._col2} + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Position of Big Table: 1 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col2 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col2 (type: int) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: boolean) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col3} + 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int), _col6 (type: int), _col7 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5 + columns.types int:int:int:int:int:int + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 50 100 50 NULL NULL +NULL NULL 100 40 NULL NULL +NULL NULL 100 60 100 60 +PREHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 50 100 50 NULL NULL +NULL NULL 100 40 NULL NULL +NULL NULL 100 60 100 60 +PREHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain extended select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t0`.`key`, `t0`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2` +FROM (SELECT `key`, CAST(50 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 50 AND `key` IS NOT NULL) AS `t0` +RIGHT JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` +FROM `default`.`a_n4`) AS `t1` ON `t0`.`key` = `t1`.`key` AND `t1`.`=3` +LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t1`.`key` = `t3`.`key` AND `t1`.`=` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n4 + filterExpr: (((value = 50) and key is not null) or ((value = 60) and key is not null)) (type: boolean) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 50 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 60) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 60 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [a_n4] + Map 4 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + filter mappings: + 1 [0, 1] + filter predicates: + 0 + 1 {VALUE._col2} + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Position of Big Table: 1 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col2 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col2 (type: int) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col0 (type: int), _col1 (type: int), _col3 (type: int), _col4 (type: boolean) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col3} + 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col6, _col7 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: int), _col6 (type: int), _col7 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5 + columns.types int:int:int:int:int:int + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 50 100 50 NULL NULL +NULL NULL 100 40 NULL NULL +NULL NULL 100 60 100 60 +PREHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ MAPJOIN(a_n4,c)*/ * from a_n4 right outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50 AND b.value>10) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND b.value>20 AND c.value=60) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 50 100 50 NULL NULL +NULL NULL 100 40 NULL NULL +NULL NULL 100 60 100 60 +PREHOOK: query: explain extended select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain extended select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t0`.`key` AS `key1`, `t0`.`value` AS `value1`, `t2`.`key` AS `key2`, `t2`.`value` AS `value2`, `t4`.`key` AS `key3`, `t4`.`value` AS `value3` +FROM (SELECT `key`, `value`, `value` = 40 AS `=`, `value` = 50 AS `=3` +FROM `default`.`a_n4`) AS `t` +FULL JOIN (SELECT `key`, `value`, `value` = 60 AS `=`, `value` = 50 AS `=3` +FROM `default`.`a_n4`) AS `t0` ON `t`.`key` = `t0`.`key` AND `t`.`=3` AND `t0`.`=3` +LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 60 AND `key` IS NOT NULL) AS `t2` ON `t0`.`key` = `t2`.`key` AND `t0`.`=` +LEFT JOIN (SELECT `key`, CAST(40 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 40 AND `key` IS NOT NULL) AS `t4` ON `t`.`key` = `t4`.`key` AND `t`.`=` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n4 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 40) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 60) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 60 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 40) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 40 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [a_n4] + Map 5 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 60) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 48 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Full Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + 1 [0, 1] + filter predicates: + 0 {VALUE._col2} + 1 {VALUE._col2} + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col6 + Position of Big Table: 0 + Statistics: Num rows: 9 Data size: 216 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col4 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col4 (type: int) + Statistics: Num rows: 9 Data size: 216 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col0 (type: int), _col1 (type: int), _col2 (type: boolean), _col5 (type: int), _col6 (type: boolean) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col5} + 1 + keys: + 0 _col4 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col4, _col5, _col8, _col9 + Position of Big Table: 0 + Statistics: Num rows: 9 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 9 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col4 (type: int), _col5 (type: int), _col8 (type: int), _col9 (type: int) + auto parallelism: true + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col1} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col4, _col5, _col8, _col9, _col10, _col11 + Position of Big Table: 0 + Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col5 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: int), _col11 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 9 Data size: 288 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 + columns.types int:int:int:int:int:int:int:int + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select * from a_n4 full outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (b.key=c.key AND b.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 40 NULL NULL NULL NULL 100 40 +100 50 100 50 NULL NULL NULL NULL +100 60 NULL NULL NULL NULL NULL NULL +NULL NULL 100 40 NULL NULL NULL NULL +NULL NULL 100 60 100 60 NULL NULL +PREHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: explain extended select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT `t`.`key`, `t`.`value`, `t1`.`key` AS `key1`, `t1`.`value` AS `value1`, `t3`.`key` AS `key2`, `t3`.`value` AS `value2`, `t5`.`key` AS `key3`, `t5`.`value` AS `value3` +FROM (SELECT `key`, `value`, `value` = 40 AS `=`, `value` = 60 AS `=3`, `value` = 50 AS `=4` +FROM `default`.`a_n4`) AS `t` +LEFT JOIN (SELECT `key`, CAST(50 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 50 AND `key` IS NOT NULL) AS `t1` ON `t`.`key` = `t1`.`key` AND `t`.`=4` +LEFT JOIN (SELECT `key`, CAST(60 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 60 AND `key` IS NOT NULL) AS `t3` ON `t`.`key` = `t3`.`key` AND `t`.`=3` +LEFT JOIN (SELECT `key`, CAST(40 AS INTEGER) AS `value` +FROM `default`.`a_n4` +WHERE `value` = 40 AND `key` IS NOT NULL) AS `t5` ON `t`.`key` = `t5`.`key` AND `t`.`=` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n4 + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: int), value (type: int), (value = 40) (type: boolean), (value = 60) (type: boolean), (value = 50) (type: boolean) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 60 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean), _col4 (type: boolean) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 60) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 60 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: ((value = 40) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 40 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [a_n4] + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 3 Data size: 24 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((value = 50) and key is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), 50 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: int) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: a_n4 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types int:int +#### A masked pattern was here #### + name default.a_n4 + numFiles 1 + numRows 3 + rawDataSize 18 + serialization.ddl struct a_n4 { i32 key, i32 value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 21 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.a_n4 + name: default.a_n4 + Truncated Path -> Alias: + /a_n4 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col3} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col5, _col6 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 72 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col3 (type: boolean), _col5 (type: int), _col6 (type: int) + auto parallelism: true + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col2} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col5, _col6, _col7, _col8 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: int) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 3 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: int), _col2 (type: boolean), _col5 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int) + auto parallelism: true + Reducer 4 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter mappings: + 0 [1, 1] + filter predicates: + 0 {VALUE._col1} + 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8, _col9, _col10 + Position of Big Table: 0 + Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int), _col9 (type: int), _col10 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 3 Data size: 96 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6,_col7 + columns.types int:int:int:int:int:int:int:int + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 40 NULL NULL NULL NULL 100 40 +100 50 100 50 NULL NULL NULL NULL +100 60 NULL NULL 100 60 NULL NULL +PREHOOK: query: select /*+ MAPJOIN(b,c, d)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n4 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ MAPJOIN(b,c, d)*/ * from a_n4 left outer join a_n4 b on (a_n4.key=b.key AND a_n4.value=50 AND b.value=50) left outer join a_n4 c on (a_n4.key=c.key AND a_n4.value=60 AND c.value=60) left outer join a_n4 d on (a_n4.key=d.key AND a_n4.value=40 AND d.value=40) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n4 +#### A masked pattern was here #### +100 40 NULL NULL NULL NULL 100 40 +100 50 100 50 NULL NULL NULL NULL +100 60 NULL NULL 100 60 NULL NULL diff --git ql/src/test/results/clientpositive/llap/join_grp_diff_keys.q.out ql/src/test/results/clientpositive/llap/join_grp_diff_keys.q.out new file mode 100644 index 0000000000..960c449f0f --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_grp_diff_keys.q.out @@ -0,0 +1,271 @@ +PREHOOK: query: create table split (id int, line_id int, orders string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@split +POSTHOOK: query: create table split (id int, line_id int, orders string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@split +PREHOOK: query: create table bar (id int, line_id int, orders string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@bar +POSTHOOK: query: create table bar (id int, line_id int, orders string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@bar +PREHOOK: query: create table foo (id int, line_id int, orders string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@foo +POSTHOOK: query: create table foo (id int, line_id int, orders string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@foo +PREHOOK: query: create table forecast (id int, line_id int, orders string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@forecast +POSTHOOK: query: create table forecast (id int, line_id int, orders string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@forecast +PREHOOK: query: explain +SELECT foo.id, count(*) as factor from + foo JOIN bar ON (foo.id = bar.id and foo.line_id = bar.line_id) + JOIN split ON (foo.id = split.id and foo.line_id = split.line_id) + JOIN forecast ON (foo.id = forecast.id AND foo.line_id = forecast.line_id) + WHERE foo.orders != 'blah' + group by foo.id +PREHOOK: type: QUERY +PREHOOK: Input: default@bar +PREHOOK: Input: default@foo +PREHOOK: Input: default@forecast +PREHOOK: Input: default@split +#### A masked pattern was here #### +POSTHOOK: query: explain +SELECT foo.id, count(*) as factor from + foo JOIN bar ON (foo.id = bar.id and foo.line_id = bar.line_id) + JOIN split ON (foo.id = split.id and foo.line_id = split.line_id) + JOIN forecast ON (foo.id = forecast.id AND foo.line_id = forecast.line_id) + WHERE foo.orders != 'blah' + group by foo.id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@bar +POSTHOOK: Input: default@foo +POSTHOOK: Input: default@forecast +POSTHOOK: Input: default@split +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) + Reducer 3 <- Map 7 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 8 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: foo + filterExpr: ((orders <> 'blah') and id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((orders <> 'blah') and id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), line_id (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: bar + filterExpr: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), line_id (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: split + filterExpr: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), line_id (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 8 + Map Operator Tree: + TableScan + alias: forecast + filterExpr: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (id is not null and line_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), line_id (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: int) + 1 _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: int) + 1 _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int), _col1 (type: int) + 1 _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count() + keys: _col0 (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: bigint) + Reducer 5 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 255 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: drop table split +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@split +PREHOOK: Output: default@split +POSTHOOK: query: drop table split +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@split +POSTHOOK: Output: default@split +PREHOOK: query: drop table bar +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@bar +PREHOOK: Output: default@bar +POSTHOOK: query: drop table bar +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@bar +POSTHOOK: Output: default@bar +PREHOOK: query: drop table foo +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@foo +PREHOOK: Output: default@foo +POSTHOOK: query: drop table foo +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@foo +POSTHOOK: Output: default@foo +PREHOOK: query: drop table forecast +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@forecast +PREHOOK: Output: default@forecast +POSTHOOK: query: drop table forecast +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@forecast +POSTHOOK: Output: default@forecast diff --git ql/src/test/results/clientpositive/llap/join_hive_626.q.out ql/src/test/results/clientpositive/llap/join_hive_626.q.out new file mode 100644 index 0000000000..acce2cd37f --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_hive_626.q.out @@ -0,0 +1,215 @@ +PREHOOK: query: create table hive_foo (foo_id int, foo_name string, foo_a string, foo_b string, +foo_c string, foo_d string) row format delimited fields terminated by ',' +stored as textfile +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@hive_foo +POSTHOOK: query: create table hive_foo (foo_id int, foo_name string, foo_a string, foo_b string, +foo_c string, foo_d string) row format delimited fields terminated by ',' +stored as textfile +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@hive_foo +PREHOOK: query: create table hive_bar (bar_id int, bar_0 int, foo_id int, bar_1 int, bar_name +string, bar_a string, bar_b string, bar_c string, bar_d string) row format +delimited fields terminated by ',' stored as textfile +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@hive_bar +POSTHOOK: query: create table hive_bar (bar_id int, bar_0 int, foo_id int, bar_1 int, bar_name +string, bar_a string, bar_b string, bar_c string, bar_d string) row format +delimited fields terminated by ',' stored as textfile +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@hive_bar +PREHOOK: query: create table hive_count (bar_id int, n int) row format delimited fields +terminated by ',' stored as textfile +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@hive_count +POSTHOOK: query: create table hive_count (bar_id int, n int) row format delimited fields +terminated by ',' stored as textfile +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@hive_count +PREHOOK: query: load data local inpath '../../data/files/hive_626_foo.txt' overwrite into table hive_foo +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@hive_foo +POSTHOOK: query: load data local inpath '../../data/files/hive_626_foo.txt' overwrite into table hive_foo +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@hive_foo +PREHOOK: query: load data local inpath '../../data/files/hive_626_bar.txt' overwrite into table hive_bar +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@hive_bar +POSTHOOK: query: load data local inpath '../../data/files/hive_626_bar.txt' overwrite into table hive_bar +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@hive_bar +PREHOOK: query: load data local inpath '../../data/files/hive_626_count.txt' overwrite into table hive_count +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@hive_count +POSTHOOK: query: load data local inpath '../../data/files/hive_626_count.txt' overwrite into table hive_count +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@hive_count +PREHOOK: query: explain +select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = +hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id +PREHOOK: type: QUERY +PREHOOK: Input: default@hive_bar +PREHOOK: Input: default@hive_count +PREHOOK: Input: default@hive_foo +#### A masked pattern was here #### +POSTHOOK: query: explain +select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = +hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@hive_bar +POSTHOOK: Input: default@hive_count +POSTHOOK: Input: default@hive_foo +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: hive_bar + filterExpr: (foo_id is not null and bar_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (foo_id is not null and bar_id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: bar_id (type: int), foo_id (type: int), bar_name (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: hive_foo + filterExpr: foo_id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: foo_id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: foo_id (type: int), foo_name (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: hive_count + filterExpr: bar_id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: bar_id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: bar_id (type: int), n (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col2, _col4 + Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 211 Basic stats: COMPLETE Column stats: NONE + value expressions: _col2 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col2, _col4, _col6 + Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col4 (type: string), _col2 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 232 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = +hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id +PREHOOK: type: QUERY +PREHOOK: Input: default@hive_bar +PREHOOK: Input: default@hive_count +PREHOOK: Input: default@hive_foo +#### A masked pattern was here #### +POSTHOOK: query: select hive_foo.foo_name, hive_bar.bar_name, n from hive_foo join hive_bar on hive_foo.foo_id = +hive_bar.foo_id join hive_count on hive_count.bar_id = hive_bar.bar_id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@hive_bar +POSTHOOK: Input: default@hive_count +POSTHOOK: Input: default@hive_foo +#### A masked pattern was here #### +foo1 bar10 2 diff --git ql/src/test/results/clientpositive/llap/join_merge_multi_expressions.q.out ql/src/test/results/clientpositive/llap/join_merge_multi_expressions.q.out new file mode 100644 index 0000000000..9c71e341b5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_merge_multi_expressions.q.out @@ -0,0 +1,188 @@ +PREHOOK: query: explain +select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain +select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), hr (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), hr (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: c + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), hr (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2000 Data size: 542000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col1 (type: string) + 1 _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12658 Data size: 3430318 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 12658 Data size: 3430318 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 12658 Data size: 2329072 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Dynamic Partitioning Event Operator + Target column: hr (string) + Target Input: c + Partition key expr: hr + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Target Vertex: Map 6 + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col1 (type: string) + 1 _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 80113 Data size: 640904 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count() + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: bigint) + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from srcpart a join srcpart b on a.key = b.key and a.hr = b.hr join srcpart c on a.hr = c.hr and a.key = c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +42464 diff --git ql/src/test/results/clientpositive/llap/join_merging.q.out ql/src/test/results/clientpositive/llap/join_merging.q.out new file mode 100644 index 0000000000..f32a740df0 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_merging.q.out @@ -0,0 +1,279 @@ +PREHOOK: query: explain select p1.p_size, p2.p_size +from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey + right outer join part p3 on p2.p_partkey = p3.p_partkey and + p1.p_size > 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_size, p2.p_size +from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey + right outer join part p3 on p2.p_partkey = p3.p_partkey and + p1.p_size > 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: p_partkey is not null (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_size (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: ((p_size > 10) and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((p_size > 10) and p_partkey is not null) (type: boolean) + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_size (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col1 (type: int) + outputColumnNames: _col1, _col3 + Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: int), _col3 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 29 Data size: 176 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3 + Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col3 (type: int), _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 21 Data size: 252 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select p1.p_size, p2.p_size +from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey + right outer join part p3 on p2.p_partkey = p3.p_partkey and + p1.p_size > 10 and p1.p_size > p2.p_size + 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@part +#### A masked pattern was here #### +POSTHOOK: query: explain select p1.p_size, p2.p_size +from part p1 left outer join part p2 on p1.p_partkey = p2.p_partkey + right outer join part p3 on p2.p_partkey = p3.p_partkey and + p1.p_size > 10 and p1.p_size > p2.p_size + 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@part +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p3 + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 104 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: (p_partkey is not null and p_size is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (p_partkey is not null and p_size is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_size (type: int), (p_size + 10) (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 26 Data size: 312 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 26 Data size: 312 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int), _col2 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: ((p_size > 10) and p_partkey is not null) (type: boolean) + Statistics: Num rows: 26 Data size: 208 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((p_size > 10) and p_partkey is not null) (type: boolean) + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: p_partkey (type: int), p_size (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 21 Data size: 168 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col1 (type: int) + outputColumnNames: _col1, _col3 + Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: int), _col3 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 33 Data size: 64 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col4 + residual filter predicates: {(_col4 > _col2)} + Statistics: Num rows: 7 Data size: 112 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col4 (type: int), _col0 (type: int), _col1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 7 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 7 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col2 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/join_on_varchar.q.out ql/src/test/results/clientpositive/llap/join_on_varchar.q.out new file mode 100644 index 0000000000..20d8b0ffcf --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_on_varchar.q.out @@ -0,0 +1,162 @@ +PREHOOK: query: create table tbl1_n3(c1 varchar(10), intcol int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tbl1_n3 +POSTHOOK: query: create table tbl1_n3(c1 varchar(10), intcol int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tbl1_n3 +PREHOOK: query: create table tbl2_n2(c2 varchar(30)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@tbl2_n2 +POSTHOOK: query: create table tbl2_n2(c2 varchar(30)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@tbl2_n2 +PREHOOK: query: insert into table tbl1_n3 select repeat('t', 10), 11 from src limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tbl1_n3 +POSTHOOK: query: insert into table tbl1_n3 select repeat('t', 10), 11 from src limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tbl1_n3 +POSTHOOK: Lineage: tbl1_n3.c1 EXPRESSION [] +POSTHOOK: Lineage: tbl1_n3.intcol SIMPLE [] +PREHOOK: query: insert into table tbl1_n3 select repeat('s', 10), 22 from src limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tbl1_n3 +POSTHOOK: query: insert into table tbl1_n3 select repeat('s', 10), 22 from src limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tbl1_n3 +POSTHOOK: Lineage: tbl1_n3.c1 EXPRESSION [] +POSTHOOK: Lineage: tbl1_n3.intcol SIMPLE [] +PREHOOK: query: insert into table tbl2_n2 select concat(repeat('t', 10), 'ppp') from src limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tbl2_n2 +POSTHOOK: query: insert into table tbl2_n2 select concat(repeat('t', 10), 'ppp') from src limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tbl2_n2 +POSTHOOK: Lineage: tbl2_n2.c2 EXPRESSION [] +PREHOOK: query: insert into table tbl2_n2 select repeat('s', 10) from src limit 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@tbl2_n2 +POSTHOOK: query: insert into table tbl2_n2 select repeat('s', 10) from src limit 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@tbl2_n2 +POSTHOOK: Lineage: tbl2_n2.c2 EXPRESSION [] +PREHOOK: query: explain +select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 +PREHOOK: type: QUERY +PREHOOK: Input: default@tbl1_n3 +PREHOOK: Input: default@tbl2_n2 +#### A masked pattern was here #### +POSTHOOK: query: explain +select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tbl1_n3 +POSTHOOK: Input: default@tbl2_n2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 2 <- Map 1 (BROADCAST_EDGE) + Reducer 3 <- Map 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: tbl1_n3 + filterExpr: c1 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: c1 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: c1 (type: varchar(10)) + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: varchar(30)) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: varchar(30)) + Statistics: Num rows: 2 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: tbl2_n2 + filterExpr: c2 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: c2 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: c2 (type: varchar(30)) + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 194 Basic stats: COMPLETE Column stats: COMPLETE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: varchar(30)) + 1 _col0 (type: varchar(30)) + outputColumnNames: _col0, _col1 + input vertices: + 0 Map 1 + Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: varchar(10)), _col1 (type: varchar(30)) + null sort order: zz + sort order: ++ + Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: varchar(10)), KEY.reducesinkkey1 (type: varchar(30)) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 382 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 +PREHOOK: type: QUERY +PREHOOK: Input: default@tbl1_n3 +PREHOOK: Input: default@tbl2_n2 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ MAPJOIN(tbl2_n2) */ c1,c2 from tbl1_n3 join tbl2_n2 on (c1 = c2) order by c1,c2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@tbl1_n3 +POSTHOOK: Input: default@tbl2_n2 +#### A masked pattern was here #### +ssssssssss ssssssssss diff --git ql/src/test/results/clientpositive/llap/join_parse.q.out ql/src/test/results/clientpositive/llap/join_parse.q.out new file mode 100644 index 0000000000..8f8ea12d9c --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_parse.q.out @@ -0,0 +1,619 @@ +PREHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key)) +inner join src src1 on src1.value =srcpart.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key)) +inner join src src1 on src1.value =srcpart.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +select srcpart.key, src1.value from +(srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain +select srcpart.key, src1.value from +(srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value) +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: value is not null (type: boolean) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 45500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col3 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 5153 Data size: 917234 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value) +inner join src src2 on src2.key = src1.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain +select srcpart.key, src1.value from +((srcpart inner join src on srcpart.key = src.key) +inner join src src1 on src1.value =srcpart.value) +inner join src src2 on src2.key = src1.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 5 (SIMPLE_EDGE) + Reducer 5 <- Map 4 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and value is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: src1 + filterExpr: (value is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (value is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 3164 Data size: 563192 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col4 + Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col4 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 8152 Data size: 1451056 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1 + Statistics: Num rows: 791 Data size: 71981 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 791 Data size: 71981 Basic stats: COMPLETE Column stats: COMPLETE + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/join_rc.q.out ql/src/test/results/clientpositive/llap/join_rc.q.out similarity index 80% rename from ql/src/test/results/clientpositive/join_rc.q.out rename to ql/src/test/results/clientpositive/llap/join_rc.q.out index d8cda5a5f2..2b136152a7 100644 --- ql/src/test/results/clientpositive/join_rc.q.out +++ ql/src/test/results/clientpositive/llap/join_rc.q.out @@ -54,63 +54,77 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: join_rc1 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: join_rc1 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: join_rc2 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: join_rc2 + filterExpr: key is not null (type: boolean) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 791 Data size: 140798 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/join_reorder.q.out ql/src/test/results/clientpositive/llap/join_reorder.q.out new file mode 100644 index 0000000000..627c492823 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_reorder.q.out @@ -0,0 +1,806 @@ +PREHOOK: query: CREATE TABLE T1_n37(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n37 +POSTHOOK: query: CREATE TABLE T1_n37(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n37 +PREHOOK: query: CREATE TABLE T2_n24(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n24 +POSTHOOK: query: CREATE TABLE T2_n24(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n24 +PREHOOK: query: CREATE TABLE T3_n8(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n8 +POSTHOOK: query: CREATE TABLE T3_n8(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n8 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n37 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n37 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n37 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n37 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n24 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t2_n24 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n24 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t2_n24 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n8 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t3_n8 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n8 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t3_n8 +PREHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT a.key, a.val, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT a.key, a.val, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(key) (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string), val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: c + filterExpr: (key + 1) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key + 1) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: (key + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (key + 1) (type: double) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double) + 1 (key + 1) (type: double) + outputColumnNames: _col0, _col1, _col5 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(key) (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string), val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: c + filterExpr: (key + 1) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key + 1) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: (key + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (key + 1) (type: double) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double) + 1 (key + 1) (type: double) + outputColumnNames: _col0, _col1, _col5 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT a.key, a.val, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +POSTHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT a.key, a.val, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +1 11 0 +1 11 0 +1 11 0 +3 13 2 +PREHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +POSTHOOK: query: FROM T1_n37 a JOIN src c ON c.key+1=a.key +SELECT /*+ STREAMTABLE(a) */ a.key, a.val, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@t1_n37 +#### A masked pattern was here #### +1 11 0 +1 11 0 +1 11 0 +3 13 2 +PREHOOK: query: EXPLAIN FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT a.key, b.key, a.val, c.val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT a.key, b.key, a.val, c.val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: c + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: val (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col5 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 val (type: string) + outputColumnNames: _col0, _col1, _col5, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col5 (type: string), _col1 (type: string), _col11 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: c + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: val (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col5 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 val (type: string) + outputColumnNames: _col0, _col1, _col5, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col5 (type: string), _col1 (type: string), _col11 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT a.key, b.key, a.val, c.val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT a.key, b.key, a.val, c.val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +2 2 12 12 +7 NULL 17 17 +NULL NULL NULL 14 +NULL NULL NULL 16 +PREHOOK: query: FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: FROM T1_n37 a + LEFT OUTER JOIN T2_n24 b ON (b.key=a.key) + RIGHT OUTER JOIN T3_n8 c ON (c.val = a.val) +SELECT /*+ STREAMTABLE(a) */ a.key, b.key, a.val, c.val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +2 2 12 12 +7 NULL 17 17 +NULL NULL NULL 14 +NULL NULL NULL 16 +PREHOOK: query: EXPLAIN FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT a.key, b.key, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT a.key, b.key, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: c + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Unique Join 0 to 0 + Unique Join 0 to 0 + Unique Join 0 to 0 + keys: + 0 key (type: string), val (type: string) + 1 key (type: string), val (type: string) + 2 key (type: string), val (type: string) + outputColumnNames: _col0, _col5, _col10 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col5 (type: string), _col10 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: c + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string), val (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: key (type: string), val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Unique Join 0 to 0 + Unique Join 0 to 0 + Unique Join 0 to 0 + keys: + 0 key (type: string), val (type: string) + 1 key (type: string), val (type: string) + 2 key (type: string), val (type: string) + outputColumnNames: _col0, _col5, _col10 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col5 (type: string), _col10 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 809 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT a.key, b.key, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT a.key, b.key, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +1 NULL NULL +2 NULL 2 +3 3 NULL +7 NULL 7 +8 8 NULL +8 8 NULL +8 NULL NULL +NULL 2 NULL +NULL 4 4 +NULL 5 NULL +NULL NULL 6 +PREHOOK: query: FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n37 +PREHOOK: Input: default@t2_n24 +PREHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +POSTHOOK: query: FROM UNIQUEJOIN + PRESERVE T1_n37 a (a.key, a.val), + PRESERVE T2_n24 b (b.key, b.val), + PRESERVE T3_n8 c (c.key, c.val) +SELECT /*+ STREAMTABLE(b) */ a.key, b.key, c.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n37 +POSTHOOK: Input: default@t2_n24 +POSTHOOK: Input: default@t3_n8 +#### A masked pattern was here #### +1 NULL NULL +2 NULL 2 +3 3 NULL +7 NULL 7 +8 8 NULL +8 8 NULL +8 NULL NULL +NULL 2 NULL +NULL 4 4 +NULL 5 NULL +NULL NULL 6 diff --git ql/src/test/results/clientpositive/llap/join_reorder2.q.out ql/src/test/results/clientpositive/llap/join_reorder2.q.out new file mode 100644 index 0000000000..6b6e081ad2 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_reorder2.q.out @@ -0,0 +1,452 @@ +PREHOOK: query: CREATE TABLE T1_n49(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n49 +POSTHOOK: query: CREATE TABLE T1_n49(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n49 +PREHOOK: query: CREATE TABLE T2_n30(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n30 +POSTHOOK: query: CREATE TABLE T2_n30(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n30 +PREHOOK: query: CREATE TABLE T3_n10(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n10 +POSTHOOK: query: CREATE TABLE T3_n10(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n10 +PREHOOK: query: CREATE TABLE T4_n3(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T4_n3 +POSTHOOK: query: CREATE TABLE T4_n3(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T4_n3 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n49 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n49 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n49 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n49 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n30 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t2_n30 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n30 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t2_n30 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n10 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t3_n10 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n10 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t3_n10 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n3 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t4_n3 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n3 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t4_n3 +PREHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON b.key = c.key + JOIN T4_n3 d ON c.key = d.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n49 +PREHOOK: Input: default@t2_n30 +PREHOOK: Input: default@t3_n10 +PREHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON b.key = c.key + JOIN T4_n3 d ON c.key = d.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n49 +POSTHOOK: Input: default@t2_n30 +POSTHOOK: Input: default@t3_n10 +POSTHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 7 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: c + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: d + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col5 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col5 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col5 (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col11 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON b.key = c.key + JOIN T4_n3 d ON c.key = d.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n49 +PREHOOK: Input: default@t2_n30 +PREHOOK: Input: default@t3_n10 +PREHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON b.key = c.key + JOIN T4_n3 d ON c.key = d.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n49 +POSTHOOK: Input: default@t2_n30 +POSTHOOK: Input: default@t3_n10 +POSTHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +2 12 2 22 2 12 2 12 +PREHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON a.val = c.val + JOIN T4_n3 d ON a.key + 1 = d.key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n49 +PREHOOK: Input: default@t2_n30 +PREHOOK: Input: default@t3_n10 +PREHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON a.val = c.val + JOIN T4_n3 d ON a.key + 1 = d.key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n49 +POSTHOOK: Input: default@t2_n30 +POSTHOOK: Input: default@t3_n10 +POSTHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 7 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: (key is not null and val is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key is not null and val is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: (key is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: c + filterExpr: val is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: d + filterExpr: (key + 1) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key + 1) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: (key + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (key + 1) (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string), val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col5 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 val (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: (_col0 + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (_col0 + 1) (type: double) + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 (_col0 + 1) (type: double) + 1 (key + 1) (type: double) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON a.val = c.val + JOIN T4_n3 d ON a.key + 1 = d.key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n49 +PREHOOK: Input: default@t2_n30 +PREHOOK: Input: default@t3_n10 +PREHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +POSTHOOK: query: SELECT /*+ STREAMTABLE(a) */ * +FROM T1_n49 a JOIN T2_n30 b ON a.key = b.key + JOIN T3_n10 c ON a.val = c.val + JOIN T4_n3 d ON a.key + 1 = d.key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n49 +POSTHOOK: Input: default@t2_n30 +POSTHOOK: Input: default@t3_n10 +POSTHOOK: Input: default@t4_n3 +#### A masked pattern was here #### +2 12 2 22 2 12 2 12 diff --git ql/src/test/results/clientpositive/llap/join_reorder3.q.out ql/src/test/results/clientpositive/llap/join_reorder3.q.out new file mode 100644 index 0000000000..13ac0dd7ac --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_reorder3.q.out @@ -0,0 +1,452 @@ +PREHOOK: query: CREATE TABLE T1_n92(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n92 +POSTHOOK: query: CREATE TABLE T1_n92(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n92 +PREHOOK: query: CREATE TABLE T2_n57(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n57 +POSTHOOK: query: CREATE TABLE T2_n57(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n57 +PREHOOK: query: CREATE TABLE T3_n21(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n21 +POSTHOOK: query: CREATE TABLE T3_n21(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n21 +PREHOOK: query: CREATE TABLE T4_n10(key STRING, val STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T4_n10 +POSTHOOK: query: CREATE TABLE T4_n10(key STRING, val STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T4_n10 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n92 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n92 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n92 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n92 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n57 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t2_n57 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n57 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t2_n57 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n21 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t3_n21 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n21 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t3_n21 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n10 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t4_n10 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T4_n10 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t4_n10 +PREHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON b.key = c.key + JOIN T4_n10 d ON c.key = d.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n92 +PREHOOK: Input: default@t2_n57 +PREHOOK: Input: default@t3_n21 +PREHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON b.key = c.key + JOIN T4_n10 d ON c.key = d.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n92 +POSTHOOK: Input: default@t2_n57 +POSTHOOK: Input: default@t3_n21 +POSTHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 7 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: c + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: d + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col5 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col5 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col5 (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col10 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col10 (type: string) + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col11 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col10 (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON b.key = c.key + JOIN T4_n10 d ON c.key = d.key +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n92 +PREHOOK: Input: default@t2_n57 +PREHOOK: Input: default@t3_n21 +PREHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON b.key = c.key + JOIN T4_n10 d ON c.key = d.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n92 +POSTHOOK: Input: default@t2_n57 +POSTHOOK: Input: default@t3_n21 +POSTHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +2 12 2 22 2 12 2 12 +PREHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON a.val = c.val + JOIN T4_n10 d ON a.key + 1 = d.key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n92 +PREHOOK: Input: default@t2_n57 +PREHOOK: Input: default@t3_n21 +PREHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON a.val = c.val + JOIN T4_n10 d ON a.key + 1 = d.key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n92 +POSTHOOK: Input: default@t2_n57 +POSTHOOK: Input: default@t3_n21 +POSTHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Map 6 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 7 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: (key is not null and val is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key is not null and val is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: b + filterExpr: (key is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key is not null and (key + 1) is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: c + filterExpr: val is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: val (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: val (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: d + filterExpr: (key + 1) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (key + 1) is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: (key + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (key + 1) (type: double) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: key (type: string), val (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col5 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: string) + 1 val (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: (_col0 + 1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: (_col0 + 1) (type: double) + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 (_col0 + 1) (type: double) + 1 (key + 1) (type: double) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11, _col15, _col16 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string), _col15 (type: string), _col16 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 488 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON a.val = c.val + JOIN T4_n10 d ON a.key + 1 = d.key + 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n92 +PREHOOK: Input: default@t2_n57 +PREHOOK: Input: default@t3_n21 +PREHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +POSTHOOK: query: SELECT /*+ STREAMTABLE(a,c) */ * +FROM T1_n92 a JOIN T2_n57 b ON a.key = b.key + JOIN T3_n21 c ON a.val = c.val + JOIN T4_n10 d ON a.key + 1 = d.key + 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n92 +POSTHOOK: Input: default@t2_n57 +POSTHOOK: Input: default@t3_n21 +POSTHOOK: Input: default@t4_n10 +#### A masked pattern was here #### +2 12 2 22 2 12 2 12 diff --git ql/src/test/results/clientpositive/llap/join_reorder4.q.out ql/src/test/results/clientpositive/llap/join_reorder4.q.out new file mode 100644 index 0000000000..11526769e4 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_reorder4.q.out @@ -0,0 +1,465 @@ +PREHOOK: query: CREATE TABLE T1_n134(key1 STRING, val1 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T1_n134 +POSTHOOK: query: CREATE TABLE T1_n134(key1 STRING, val1 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T1_n134 +PREHOOK: query: CREATE TABLE T2_n80(key2 STRING, val2 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T2_n80 +POSTHOOK: query: CREATE TABLE T2_n80(key2 STRING, val2 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T2_n80 +PREHOOK: query: CREATE TABLE T3_n32(key3 STRING, val3 STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@T3_n32 +POSTHOOK: query: CREATE TABLE T3_n32(key3 STRING, val3 STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@T3_n32 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n134 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t1_n134 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T1.txt' INTO TABLE T1_n134 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t1_n134 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n80 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t2_n80 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T2.txt' INTO TABLE T2_n80 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t2_n80 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n32 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@t3_n32 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/T3.txt' INTO TABLE T3_n32 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@t3_n32 +PREHOOK: query: explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: explain select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key2 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: c + filterExpr: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key3 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key3 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key1 (type: string) + 1 key2 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 key3 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ STREAMTABLE(a) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +2 12 2 22 2 12 +PREHOOK: query: explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: explain select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key2 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: c + filterExpr: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key3 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key3 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key1 (type: string) + 1 key2 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 key3 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ STREAMTABLE(b) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +2 12 2 22 2 12 +PREHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: explain select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key1 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key2 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key2 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: c + filterExpr: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key3 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key3 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: val3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key1 (type: string) + 1 key2 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col5 (type: string), _col6 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 key3 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6, _col10, _col11 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col10 (type: string), _col11 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +PREHOOK: type: QUERY +PREHOOK: Input: default@t1_n134 +PREHOOK: Input: default@t2_n80 +PREHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +POSTHOOK: query: select /*+ STREAMTABLE(c) */ a.*, b.*, c.* from T1_n134 a join T2_n80 b on a.key1=b.key2 join T3_n32 c on a.key1=c.key3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@t1_n134 +POSTHOOK: Input: default@t2_n80 +POSTHOOK: Input: default@t3_n32 +#### A masked pattern was here #### +2 12 2 22 2 12 diff --git ql/src/test/results/clientpositive/llap/join_star.q.out ql/src/test/results/clientpositive/llap/join_star.q.out new file mode 100644 index 0000000000..16d539c4c3 --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_star.q.out @@ -0,0 +1,1011 @@ +PREHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@fact +POSTHOOK: query: create table fact(m1 int, m2 int, d1 int, d2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@fact +PREHOOK: query: create table dim1(f1 int, f2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim1 +POSTHOOK: query: create table dim1(f1 int, f2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim1 +PREHOOK: query: create table dim2(f3 int, f4 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim2 +POSTHOOK: query: create table dim2(f3 int, f4 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim2 +PREHOOK: query: create table dim3(f5 int, f6 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim3 +POSTHOOK: query: create table dim3(f5 int, f6 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim3 +PREHOOK: query: create table dim4(f7 int, f8 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim4 +POSTHOOK: query: create table dim4(f7 int, f8 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim4 +PREHOOK: query: create table dim5(f9 int, f10 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim5 +POSTHOOK: query: create table dim5(f9 int, f10 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim5 +PREHOOK: query: create table dim6(f11 int, f12 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim6 +POSTHOOK: query: create table dim6(f11 int, f12 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim6 +PREHOOK: query: create table dim7(f13 int, f14 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dim7 +POSTHOOK: query: create table dim7(f13 int, f14 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dim7 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@fact +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/fact-data.txt' INTO TABLE fact +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@fact +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim1 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim1 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim1 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim2 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim2 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim2 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim3 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim3 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim3 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim4 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim4 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim4 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim5 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim5 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim5 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim6 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim6 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim6 +PREHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7 +PREHOOK: type: LOAD +#### A masked pattern was here #### +PREHOOK: Output: default@dim7 +POSTHOOK: query: LOAD DATA LOCAL INPATH '../../data/files/dim-data.txt' INTO TABLE dim7 +POSTHOOK: type: LOAD +#### A masked pattern was here #### +POSTHOOK: Output: default@dim7 +PREHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: explain select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact + filterExpr: d1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: d1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: m1 (type: int), m2 (type: int), d1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col4 + input vertices: + 1 Map 2 + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: dim1 + filterExpr: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f1 (type: int), f2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: select m1, m2, f2 from fact join dim1 on fact.d1=dim1.f1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +1 2 1 +11 12 1 +21 22 1 +31 32 1 +41 42 1 +51 52 3 +61 62 3 +71 72 3 +81 82 3 +91 92 3 +PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact + filterExpr: (d1 is not null and d2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (d1 is not null and d2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: m1 (type: int), m2 (type: int), d1 (type: int), d2 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3, _col5 + input vertices: + 1 Map 2 + Statistics: Num rows: 1 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col3 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7 + input vertices: + 1 Map 3 + Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col7 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: dim1 + filterExpr: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f1 (type: int), f2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: dim2 + filterExpr: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f3 (type: int), f4 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1=dim1.f1 join dim2 on fact.d2=dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +1 2 1 2 +11 12 1 2 +21 22 1 2 +31 32 1 2 +41 42 1 2 +51 52 3 4 +61 62 3 4 +71 72 3 4 +81 82 3 4 +91 92 3 4 +PREHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: explain select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: dim1 + filterExpr: (f1 is not null and f2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (f1 is not null and f2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f1 (type: int), f2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3 + input vertices: + 1 Map 2 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col2 (type: int) + outputColumnNames: _col1, _col3, _col4, _col5 + input vertices: + 1 Map 3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col4 (type: int), _col5 (type: int), _col1 (type: int), _col3 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: dim2 + filterExpr: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f3 (type: int), f4 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: fact + filterExpr: d1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: d1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: m1 (type: int), m2 (type: int), d1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: int) + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: select m1, m2, f2, f4 from fact join dim1 on fact.d1= dim1.f1 join dim2 on dim1.f2 = dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +1 2 1 1 +11 12 1 1 +21 22 1 1 +31 32 1 1 +41 42 1 1 +51 52 3 3 +61 62 3 3 +71 72 3 3 +81 82 3 3 +91 92 3 3 +PREHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: explain select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: m1 (type: int), m2 (type: int), d1 (type: int) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 12 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col4 + input vertices: + 1 Map 2 + Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col4 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col4, _col6 + input vertices: + 1 Map 3 + Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col4 (type: int), _col6 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 14 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: dim1 + filterExpr: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f1 (type: int), f2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: dim2 + filterExpr: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f3 (type: int), f4 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: select m1, m2, f2, f4 from fact Left outer join dim1 on fact.d1= dim1.f1 Left outer join dim2 on dim1.f2 = dim2.f3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +1 2 1 1 +11 12 1 1 +21 22 1 1 +31 32 1 1 +41 42 1 1 +51 52 3 3 +61 62 3 3 +71 72 3 3 +81 82 3 3 +91 92 3 3 +PREHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14 + from fact + Left outer join dim1 on fact.d1= dim1.f1 + Left outer join dim2 on dim1.f2 = dim2.f3 + Left outer Join dim3 on fact.d2= dim3.f5 + Left outer Join dim4 on dim3.f6= dim4.f7 + Left outer join dim5 on dim4.f8= dim5.f9 + Left outer Join dim6 on dim3.f6= dim6.f11 + Left outer Join dim7 on dim6.f12 = dim7.f13 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@dim3 +PREHOOK: Input: default@dim4 +PREHOOK: Input: default@dim5 +PREHOOK: Input: default@dim6 +PREHOOK: Input: default@dim7 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: explain Select m1, m2, f2, f4, f6, f8, f10, f12, f14 + from fact + Left outer join dim1 on fact.d1= dim1.f1 + Left outer join dim2 on dim1.f2 = dim2.f3 + Left outer Join dim3 on fact.d2= dim3.f5 + Left outer Join dim4 on dim3.f6= dim4.f7 + Left outer join dim5 on dim4.f8= dim5.f9 + Left outer Join dim6 on dim3.f6= dim6.f11 + Left outer Join dim7 on dim6.f12 = dim7.f13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@dim3 +POSTHOOK: Input: default@dim4 +POSTHOOK: Input: default@dim5 +POSTHOOK: Input: default@dim6 +POSTHOOK: Input: default@dim7 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Map 2 (BROADCAST_EDGE), Map 3 (BROADCAST_EDGE), Map 4 (BROADCAST_EDGE), Map 5 (BROADCAST_EDGE), Map 6 (BROADCAST_EDGE), Map 7 (BROADCAST_EDGE), Map 8 (BROADCAST_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: m1 (type: int), m2 (type: int), d1 (type: int), d2 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 16 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col2 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3, _col5 + input vertices: + 1 Map 2 + Statistics: Num rows: 1 Data size: 17 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col5 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col3, _col5, _col7 + input vertices: + 1 Map 3 + Statistics: Num rows: 1 Data size: 18 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col3 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7, _col9 + input vertices: + 1 Map 4 + Statistics: Num rows: 1 Data size: 19 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col9 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11 + input vertices: + 1 Map 5 + Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col11 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13 + input vertices: + 1 Map 6 + Statistics: Num rows: 1 Data size: 22 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col9 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13, _col15 + input vertices: + 1 Map 7 + Statistics: Num rows: 1 Data size: 24 Basic stats: COMPLETE Column stats: NONE + Map Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col15 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col5, _col7, _col9, _col11, _col13, _col15, _col17 + input vertices: + 1 Map 8 + Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col5 (type: int), _col7 (type: int), _col9 (type: int), _col11 (type: int), _col13 (type: int), _col15 (type: int), _col17 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 + Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 26 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 2 + Map Operator Tree: + TableScan + alias: dim1 + filterExpr: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f1 (type: int), f2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: dim2 + filterExpr: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f3 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f3 (type: int), f4 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: dim3 + filterExpr: f5 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f5 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f5 (type: int), f6 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: dim4 + filterExpr: f7 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f7 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f7 (type: int), f8 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: dim5 + filterExpr: f9 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f9 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f9 (type: int), f10 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: dim6 + filterExpr: f11 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f11 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f11 (type: int), f12 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 8 + Map Operator Tree: + TableScan + alias: dim7 + filterExpr: f13 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: f13 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: f13 (type: int), f14 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14 + from fact + Left outer join dim1 on fact.d1= dim1.f1 + Left outer join dim2 on dim1.f2 = dim2.f3 + Left outer Join dim3 on fact.d2= dim3.f5 + Left outer Join dim4 on dim3.f6= dim4.f7 + Left outer join dim5 on dim4.f8= dim5.f9 + Left outer Join dim6 on dim3.f6= dim6.f11 + Left outer Join dim7 on dim6.f12 = dim7.f13 +PREHOOK: type: QUERY +PREHOOK: Input: default@dim1 +PREHOOK: Input: default@dim2 +PREHOOK: Input: default@dim3 +PREHOOK: Input: default@dim4 +PREHOOK: Input: default@dim5 +PREHOOK: Input: default@dim6 +PREHOOK: Input: default@dim7 +PREHOOK: Input: default@fact +#### A masked pattern was here #### +POSTHOOK: query: Select m1, m2, f2, f4, f6, f8, f10, f12, f14 + from fact + Left outer join dim1 on fact.d1= dim1.f1 + Left outer join dim2 on dim1.f2 = dim2.f3 + Left outer Join dim3 on fact.d2= dim3.f5 + Left outer Join dim4 on dim3.f6= dim4.f7 + Left outer join dim5 on dim4.f8= dim5.f9 + Left outer Join dim6 on dim3.f6= dim6.f11 + Left outer Join dim7 on dim6.f12 = dim7.f13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@dim1 +POSTHOOK: Input: default@dim2 +POSTHOOK: Input: default@dim3 +POSTHOOK: Input: default@dim4 +POSTHOOK: Input: default@dim5 +POSTHOOK: Input: default@dim6 +POSTHOOK: Input: default@dim7 +POSTHOOK: Input: default@fact +#### A masked pattern was here #### +1 2 1 1 2 2 2 2 2 +11 12 1 1 2 2 2 2 2 +21 22 1 1 2 2 2 2 2 +31 32 1 1 2 2 2 2 2 +41 42 1 1 2 2 2 2 2 +51 52 3 3 4 4 4 4 4 +61 62 3 3 4 4 4 4 4 +71 72 3 3 4 4 4 4 4 +81 82 3 3 4 4 4 4 4 +91 92 3 3 4 4 4 4 4 diff --git ql/src/test/results/clientpositive/llap/join_thrift.q.out ql/src/test/results/clientpositive/llap/join_thrift.q.out new file mode 100644 index 0000000000..63d7d4f0ee --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_thrift.q.out @@ -0,0 +1,133 @@ +PREHOOK: query: DESCRIBE src_thrift +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@src_thrift +POSTHOOK: query: DESCRIBE src_thrift +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@src_thrift +aint int from deserializer +astring string from deserializer +lint array from deserializer +lstring array from deserializer +lintstring array> from deserializer +mstringstring map from deserializer +attributes map,map>>>> from deserializer +unionfield1 uniontype,map> from deserializer +unionfield2 uniontype,map> from deserializer +unionfield3 uniontype,map> from deserializer +PREHOOK: query: EXPLAIN +SELECT s1.aint, s2.lintstring +FROM src_thrift s1 +JOIN src_thrift s2 +ON s1.aint = s2.aint +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT s1.aint, s2.lintstring +FROM src_thrift s1 +JOIN src_thrift s2 +ON s1.aint = s2.aint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: s1 + filterExpr: aint is not null (type: boolean) + Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: aint is not null (type: boolean) + Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: aint (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: aint (type: int) + Statistics: Num rows: 11 Data size: 44 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: s2 + filterExpr: aint is not null (type: boolean) + Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: aint is not null (type: boolean) + Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: aint (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: aint (type: int) + Statistics: Num rows: 11 Data size: 28204 Basic stats: COMPLETE Column stats: NONE + value expressions: lintstring (type: array>) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 aint (type: int) + 1 aint (type: int) + outputColumnNames: _col0, _col17 + Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: int), _col17 (type: array>) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 12 Data size: 48 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT s1.aint, s2.lintstring +FROM src_thrift s1 +JOIN src_thrift s2 +ON s1.aint = s2.aint +PREHOOK: type: QUERY +PREHOOK: Input: default@src_thrift +#### A masked pattern was here #### +POSTHOOK: query: SELECT s1.aint, s2.lintstring +FROM src_thrift s1 +JOIN src_thrift s2 +ON s1.aint = s2.aint +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src_thrift +#### A masked pattern was here #### +-1461153973 [{"myint":49,"mystring":"343","underscore_int":7}] +-1952710710 [{"myint":25,"mystring":"125","underscore_int":5}] +-734328909 [{"myint":16,"mystring":"64","underscore_int":4}] +-751827638 [{"myint":4,"mystring":"8","underscore_int":2}] +0 NULL +1244525190 [{"myint":36,"mystring":"216","underscore_int":6}] +1638581578 [{"myint":64,"mystring":"512","underscore_int":8}] +1712634731 [{"myint":0,"mystring":"0","underscore_int":0}] +336964413 [{"myint":81,"mystring":"729","underscore_int":9}] +465985200 [{"myint":1,"mystring":"1","underscore_int":1}] +477111222 [{"myint":9,"mystring":"27","underscore_int":3}] diff --git ql/src/test/results/clientpositive/llap/join_view.q.out ql/src/test/results/clientpositive/llap/join_view.q.out new file mode 100644 index 0000000000..456d62559c --- /dev/null +++ ql/src/test/results/clientpositive/llap/join_view.q.out @@ -0,0 +1,161 @@ +PREHOOK: query: drop table invites +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table invites +POSTHOOK: type: DROPTABLE +PREHOOK: query: drop table invites2 +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table invites2 +POSTHOOK: type: DROPTABLE +PREHOOK: query: create table invites (foo int, bar string) partitioned by (ds string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@invites +POSTHOOK: query: create table invites (foo int, bar string) partitioned by (ds string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@invites +PREHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@invites2 +POSTHOOK: query: create table invites2 (foo int, bar string) partitioned by (ds string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@invites2 +PREHOOK: query: create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@invites +PREHOOK: Input: default@invites2 +PREHOOK: Output: database:default +PREHOOK: Output: default@v +POSTHOOK: query: create view v as select invites.bar, invites2.foo, invites2.ds from invites join invites2 on invites.ds=invites2.ds +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@invites +POSTHOOK: Input: default@invites2 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@v +POSTHOOK: Lineage: v.bar SIMPLE [(invites)invites.FieldSchema(name:bar, type:string, comment:null), ] +POSTHOOK: Lineage: v.ds SIMPLE [(invites2)invites2.FieldSchema(name:ds, type:string, comment:null), ] +POSTHOOK: Lineage: v.foo SIMPLE [(invites2)invites2.FieldSchema(name:foo, type:int, comment:null), ] +Warning: Shuffle Join MERGEJOIN[13][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: explain select * from v where ds='2011-09-01' +PREHOOK: type: QUERY +PREHOOK: Input: default@invites +PREHOOK: Input: default@invites2 +PREHOOK: Input: default@v +#### A masked pattern was here #### +POSTHOOK: query: explain select * from v where ds='2011-09-01' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@invites +POSTHOOK: Input: default@invites2 +POSTHOOK: Input: default@v +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (XPROD_EDGE), Map 3 (XPROD_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: invites + filterExpr: (ds = '2011-09-01') (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (ds = '2011-09-01') (type: boolean) + Statistics: Num rows: 1 Data size: 268 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: bar (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 84 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: unknown + Map 3 + Map Operator Tree: + TableScan + alias: invites2 + filterExpr: (ds = '2011-09-01') (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (ds = '2011-09-01') (type: boolean) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: foo (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 0 Basic stats: PARTIAL Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: unknown + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: int), '2011-09-01' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 85 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: drop view v +PREHOOK: type: DROPVIEW +PREHOOK: Input: default@v +PREHOOK: Output: default@v +POSTHOOK: query: drop view v +POSTHOOK: type: DROPVIEW +POSTHOOK: Input: default@v +POSTHOOK: Output: default@v +PREHOOK: query: drop table invites +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@invites +PREHOOK: Output: default@invites +POSTHOOK: query: drop table invites +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@invites +POSTHOOK: Output: default@invites +PREHOOK: query: drop table invites2 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@invites2 +PREHOOK: Output: default@invites2 +POSTHOOK: query: drop table invites2 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@invites2 +POSTHOOK: Output: default@invites2 diff --git ql/src/test/results/clientpositive/keyword_1.q.out ql/src/test/results/clientpositive/llap/keyword_1.q.out similarity index 90% rename from ql/src/test/results/clientpositive/keyword_1.q.out rename to ql/src/test/results/clientpositive/llap/keyword_1.q.out index 609c2457c6..33ceee58c2 100644 --- ql/src/test/results/clientpositive/keyword_1.q.out +++ ql/src/test/results/clientpositive/llap/keyword_1.q.out @@ -30,11 +30,9 @@ STAGE PLANS: Processor Tree: TableScan alias: test_user - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: user (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: show grant user hive_test on table test_user @@ -82,11 +80,9 @@ STAGE PLANS: Processor Tree: TableScan alias: test_user - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: role (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: show grant user hive_test on table test_user diff --git ql/src/test/results/clientpositive/llap/lateral_view_cp.q.out ql/src/test/results/clientpositive/llap/lateral_view_cp.q.out new file mode 100644 index 0000000000..ea2b1f056b --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_cp.q.out @@ -0,0 +1,167 @@ +PREHOOK: query: drop table array_valued_src +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table array_valued_src +POSTHOOK: type: DROPTABLE +PREHOOK: query: create table array_valued_src (key string, value array) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@array_valued_src +POSTHOOK: query: create table array_valued_src (key string, value array) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@array_valued_src +PREHOOK: query: insert overwrite table array_valued_src select key, array(value) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@array_valued_src +POSTHOOK: query: insert overwrite table array_valued_src select key, array(value) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@array_valued_src +POSTHOOK: Lineage: array_valued_src.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: array_valued_src.value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val +PREHOOK: type: QUERY +PREHOOK: Input: default@array_valued_src +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@array_valued_src +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 1001504 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 475 Data size: 951428 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 475 Data size: 951428 Basic stats: COMPLETE Column stats: NONE + value expressions: value (type: array) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 key (type: string) + outputColumnNames: _col6 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col6 (type: array) + outputColumnNames: _col1 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Lateral View Forward + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Select Operator + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + Lateral View Join Operator + outputColumnNames: _col2 + Statistics: Num rows: 1100 Data size: 95700 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(_col2) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: bigint) + Select Operator + expressions: _col1 (type: array) + outputColumnNames: _col0 + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + UDTF Operator + Statistics: Num rows: 550 Data size: 47850 Basic stats: COMPLETE Column stats: NONE + function name: explode + Lateral View Join Operator + outputColumnNames: _col2 + Statistics: Num rows: 1100 Data size: 95700 Basic stats: COMPLETE Column stats: NONE + Group By Operator + aggregations: count(_col2) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: bigint) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val +PREHOOK: type: QUERY +PREHOOK: Input: default@array_valued_src +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select count(val) from (select a.key as key, b.value as array_val from src a join array_valued_src b on a.key=b.key) i lateral view explode (array_val) c as val +POSTHOOK: type: QUERY +POSTHOOK: Input: default@array_valued_src +POSTHOOK: Input: default@src +#### A masked pattern was here #### +1028 diff --git ql/src/test/results/clientpositive/llap/lateral_view_explode2.q.out ql/src/test/results/clientpositive/llap/lateral_view_explode2.q.out new file mode 100644 index 0000000000..8f8a4cf6be --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_explode2.q.out @@ -0,0 +1,130 @@ +PREHOOK: query: CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2' +PREHOOK: type: CREATEFUNCTION +PREHOOK: Output: explode2 +POSTHOOK: query: CREATE TEMPORARY FUNCTION explode2 AS 'org.apache.hadoop.hive.contrib.udtf.example.GenericUDTFExplode2' +POSTHOOK: type: CREATEFUNCTION +POSTHOOK: Output: explode2 +PREHOOK: query: EXPLAIN SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Lateral View Forward + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: _col5 (type: int), _col6 (type: int) + null sort order: zz + Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 3 + Group By Operator + keys: _col5 (type: int), _col6 (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE + UDTF Operator + Statistics: Num rows: 500 Data size: 28000 Basic stats: COMPLETE Column stats: COMPLETE + function name: explode + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: _col5 (type: int), _col6 (type: int) + null sort order: zz + Statistics: Num rows: 1000 Data size: 200000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 3 + Group By Operator + keys: _col5 (type: int), _col6 (type: int) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: int) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 3 + Processor Tree: + ListSink + +PREHOOK: query: SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT col1, col2 FROM src LATERAL VIEW explode2(array(1,2,3)) myTable AS col1, col2 group by col1, col2 LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +1 1 +2 2 +3 3 +PREHOOK: query: DROP TEMPORARY FUNCTION explode2 +PREHOOK: type: DROPFUNCTION +PREHOOK: Output: explode2 +POSTHOOK: query: DROP TEMPORARY FUNCTION explode2 +POSTHOOK: type: DROPFUNCTION +POSTHOOK: Output: explode2 diff --git ql/src/test/results/clientpositive/llap/lateral_view_noalias.q.out ql/src/test/results/clientpositive/llap/lateral_view_noalias.q.out new file mode 100644 index 0000000000..c6a69f6946 --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_noalias.q.out @@ -0,0 +1,296 @@ +PREHOOK: query: EXPLAIN SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 2 + Processor Tree: + TableScan + alias: src + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 2 + ListSink + Select Operator + expressions: map('key1':100,'key2':200) (type: map) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 2 + ListSink + +PREHOOK: query: SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +key1 100 +key2 200 +PREHOOK: query: EXPLAIN SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 2 + Processor Tree: + TableScan + alias: src + Select Operator + expressions: map('key1':100,'key2':200) (type: map) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Select Operator + expressions: key (type: string), value (type: int) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 2 + ListSink + +PREHOOK: query: SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT explode(map('key1', 100, 'key2', 200)) from src limit 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +key1 100 +key2 200 +PREHOOK: query: create view lv_noalias as SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_noalias +POSTHOOK: query: create view lv_noalias as SELECT myTab.* from src LATERAL VIEW explode(map('key1', 100, 'key2', 200)) myTab limit 2 +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_noalias +POSTHOOK: Lineage: lv_noalias.key SCRIPT [] +POSTHOOK: Lineage: lv_noalias.value SCRIPT [] +PREHOOK: query: explain select * from lv_noalias a join lv_noalias b on a.key=b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_noalias +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from lv_noalias a join lv_noalias b on a.key=b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_noalias +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Lateral View Forward + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: int) + Select Operator + expressions: map('key1':100,'key2':200) (type: map) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE + UDTF Operator + Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE + function name: explode + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: int) + Select Operator + Statistics: Num rows: 500 Data size: 172000 Basic stats: COMPLETE Column stats: COMPLETE + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: int) + Select Operator + expressions: map('key1':100,'key2':200) (type: map) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE + UDTF Operator + Statistics: Num rows: 500 Data size: 192000 Basic stats: COMPLETE Column stats: COMPLETE + function name: explode + Lateral View Join Operator + outputColumnNames: _col5, _col6 + Statistics: Num rows: 1000 Data size: 364000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col5 (type: string), _col6 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 8000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: int) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 432 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: _col0 is not null (type: boolean) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from lv_noalias a join lv_noalias b on a.key=b.key +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_noalias +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from lv_noalias a join lv_noalias b on a.key=b.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_noalias +POSTHOOK: Input: default@src +#### A masked pattern was here #### +key1 100 key1 100 +key2 200 key2 200 diff --git ql/src/test/results/clientpositive/llap/lateral_view_onview.q.out ql/src/test/results/clientpositive/llap/lateral_view_onview.q.out new file mode 100644 index 0000000000..35845d7be5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_onview.q.out @@ -0,0 +1,711 @@ +PREHOOK: query: CREATE TABLE lv_table_n0( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_table_n0 +POSTHOOK: query: CREATE TABLE lv_table_n0( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_table_n0 +PREHOOK: query: INSERT OVERWRITE TABLE lv_table_n0 SELECT 'abc ', array(1,2,3), 100, 't' FROM src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@lv_table_n0 +POSTHOOK: query: INSERT OVERWRITE TABLE lv_table_n0 SELECT 'abc ', array(1,2,3), 100, 't' FROM src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@lv_table_n0 +POSTHOOK: Lineage: lv_table_n0.c1 SIMPLE [] +POSTHOOK: Lineage: lv_table_n0.c2 EXPRESSION [] +POSTHOOK: Lineage: lv_table_n0.c3 SIMPLE [] +POSTHOOK: Lineage: lv_table_n0.c4 EXPRESSION [] +PREHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n0 +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_view +POSTHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n0 +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_view +POSTHOOK: Lineage: lv_view.c1 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c1, type:string, comment:null), ] +POSTHOOK: Lineage: lv_view.c2 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c2, type:array, comment:null), ] +POSTHOOK: Lineage: lv_view.c3 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c3, type:int, comment:null), ] +POSTHOOK: Lineage: lv_view.c4 SIMPLE [(lv_table_n0)lv_table_n0.FieldSchema(name:c4, type:char(1), comment:null), ] +PREHOOK: query: EXPLAIN SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: lv_table_n0 + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: c1 (type: string), c2 (type: array), c3 (type: int), c4 (type: char(1)) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + Lateral View Forward + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col4 (type: int) + null sort order: zz + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col4 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + UDTF Operator + Statistics: Num rows: 500 Data size: 187068 Basic stats: COMPLETE Column stats: NONE + function name: explode + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col4 (type: int) + null sort order: zz + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col4 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: array), VALUE._col1 (type: int), VALUE._col2 (type: char(1)), KEY.reducesinkkey1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1000 Data size: 374136 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col4 (type: int) + null sort order: zz + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col4 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: array), VALUE._col1 (type: int), VALUE._col2 (type: char(1)), KEY.reducesinkkey1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 374 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: EXPLAIN SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 3 + Processor Tree: + TableScan + alias: lv_table_n0 + Select Operator + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col4 + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4 + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + +PREHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 9 + Processor Tree: + TableScan + alias: lv_table_n0 + Select Operator + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col4 + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array('a','b','c') (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col4 + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array('a','b','c') (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + +PREHOOK: query: EXPLAIN SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 3 + Processor Tree: + TableScan + alias: lv_table_n0 + Select Operator + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col5 + Select Operator + expressions: _col5 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + Select Operator + expressions: _col4 (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col5 + Select Operator + expressions: _col5 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + Select Operator + expressions: array(array(1,2,3)) (type: array>) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col5 + Select Operator + expressions: _col5 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + Select Operator + expressions: _col4 (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col5 + Select Operator + expressions: _col5 (type: int) + outputColumnNames: _col0 + Limit + Number of rows: 3 + ListSink + +PREHOOK: query: SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: SELECT * FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol SORT BY c1 ASC, myCol ASC LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +abc [1,2,3] 100 t 1 +PREHOOK: query: SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: SELECT myTable.* FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +1 +2 +3 +PREHOOK: query: SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +1 a +1 b +1 c +2 a +2 b +2 c +3 a +3 b +3 c +PREHOOK: query: SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: SELECT myTable2.* FROM lv_view LATERAL VIEW explode(array(array(1,2,3))) myTable AS myCol LATERAL VIEW explode(myTable.myCol) myTable2 AS myCol2 LIMIT 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +1 +2 +3 +PREHOOK: query: EXPLAIN +SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN +SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 3 + Processor Tree: + TableScan + alias: lv_table_n0 + Select Operator + expressions: c1 (type: string), c2 (type: array), c3 (type: int), c4 (type: char(1)) + outputColumnNames: _col0, _col1, _col2, _col3 + Lateral View Forward + Select Operator + expressions: _col0 (type: string), _col1 (type: array), _col2 (type: int), _col3 (type: char(1)) + outputColumnNames: _col0, _col1, _col2, _col3 + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Select Operator + expressions: size(_col1) (type: int), _col2 (type: int), trim(_col0) (type: string), _col3 (type: char(1)), _col4 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Limit + Number of rows: 3 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Select Operator + expressions: size(_col1) (type: int), _col2 (type: int), trim(_col0) (type: string), _col3 (type: char(1)), _col4 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Limit + Number of rows: 3 + ListSink + +PREHOOK: query: SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n0 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: SELECT SIZE(c2),c3,TRIM(c1),c4,myCol from lv_view LATERAL VIEW explode(array(1,2,3)) myTab as myCol limit 3 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n0 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +3 100 abc t 1 +3 100 abc t 2 +3 100 abc t 3 +PREHOOK: query: CREATE TABLE lv_table1_n0( c1 STRING, c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_table1_n0 +POSTHOOK: query: CREATE TABLE lv_table1_n0( c1 STRING, c3 INT, c4 CHAR(1), c5 STRING, c6 STRING, c7 STRING, c8 STRING, c9 STRING, c10 STRING, c11 STRING, c12 STRING, c13 STRING) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_table1_n0 +PREHOOK: query: CREATE TABLE lv_table2( c1 STRING, c2 ARRAY) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_table2 +POSTHOOK: query: CREATE TABLE lv_table2( c1 STRING, c2 ARRAY) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_table2 +PREHOOK: query: INSERT OVERWRITE TABLE lv_table1_n0 SELECT 'abc ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@lv_table1_n0 +POSTHOOK: query: INSERT OVERWRITE TABLE lv_table1_n0 SELECT 'abc ', 100, 't', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test', 'test' FROM src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@lv_table1_n0 +POSTHOOK: Lineage: lv_table1_n0.c1 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c10 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c11 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c12 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c13 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c3 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c4 EXPRESSION [] +POSTHOOK: Lineage: lv_table1_n0.c5 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c6 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c7 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c8 SIMPLE [] +POSTHOOK: Lineage: lv_table1_n0.c9 SIMPLE [] +PREHOOK: query: INSERT OVERWRITE TABLE lv_table2 SELECT 'abc ', array(1,2,3) FROM src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@lv_table2 +POSTHOOK: query: INSERT OVERWRITE TABLE lv_table2 SELECT 'abc ', array(1,2,3) FROM src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@lv_table2 +POSTHOOK: Lineage: lv_table2.c1 SIMPLE [] +POSTHOOK: Lineage: lv_table2.c2 EXPRESSION [] +PREHOOK: query: EXPLAIN WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table1_n0 +PREHOOK: Input: default@lv_table2 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table1_n0 +POSTHOOK: Input: default@lv_table2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: lv_table1_n0 + filterExpr: c1 is not null (type: boolean) + Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: c1 is not null (type: boolean) + Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: c1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: c1 (type: string) + Statistics: Num rows: 500 Data size: 485000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: c3 (type: int), c4 (type: char(1)), c5 (type: string), c6 (type: string), c7 (type: string), c8 (type: string), c9 (type: string), c10 (type: string), c11 (type: string), c12 (type: string), c13 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: lv_table2 + filterExpr: c1 is not null (type: boolean) + Statistics: Num rows: 500 Data size: 144704 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: c1 is not null (type: boolean) + Statistics: Num rows: 475 Data size: 137468 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: c1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: c1 (type: string) + Statistics: Num rows: 475 Data size: 137468 Basic stats: COMPLETE Column stats: NONE + value expressions: c2 (type: array) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 c1 (type: string) + 1 c1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col16 + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col16 (type: array) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + Lateral View Forward + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col13 (type: int) + null sort order: zz + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col13 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) + Select Operator + expressions: _col12 (type: array) + outputColumnNames: _col0 + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + UDTF Operator + Statistics: Num rows: 550 Data size: 533500 Basic stats: COMPLETE Column stats: NONE + function name: explode + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col13 (type: int) + null sort order: zz + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col13 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: int), VALUE._col1 (type: char(1)), VALUE._col2 (type: string), VALUE._col3 (type: string), VALUE._col4 (type: string), VALUE._col5 (type: string), VALUE._col6 (type: string), VALUE._col7 (type: string), VALUE._col8 (type: string), VALUE._col9 (type: string), VALUE._col10 (type: string), VALUE._col11 (type: array), KEY.reducesinkkey1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + Statistics: Num rows: 1100 Data size: 1067000 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + Top N Key Operator + sort order: ++ + keys: _col0 (type: string), _col13 (type: int) + null sort order: zz + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + top n: 1 + Reduce Output Operator + key expressions: _col0 (type: string), _col13 (type: int) + null sort order: zz + sort order: ++ + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + TopN Hash Memory Usage: 0.1 + value expressions: _col1 (type: int), _col2 (type: char(1)), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string), _col9 (type: string), _col10 (type: string), _col11 (type: string), _col12 (type: array) + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), VALUE._col0 (type: int), VALUE._col1 (type: char(1)), VALUE._col2 (type: string), VALUE._col3 (type: string), VALUE._col4 (type: string), VALUE._col5 (type: string), VALUE._col6 (type: string), VALUE._col7 (type: string), VALUE._col8 (type: string), VALUE._col9 (type: string), VALUE._col10 (type: string), VALUE._col11 (type: array), KEY.reducesinkkey1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12, _col13 + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 970 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table1_n0 +PREHOOK: Input: default@lv_table2 +#### A masked pattern was here #### +POSTHOOK: query: WITH lv_view1 AS (SELECT lv_table1_n0.*, c2 FROM lv_table1_n0 JOIN lv_table2 ON lv_table1_n0.c1 = lv_table2.c1), lv_view2 AS (SELECT * FROM lv_view1 LATERAL VIEW explode(c2) myTable AS myCol) SELECT * FROM lv_view2 SORT BY c1 ASC, myCol ASC LIMIT 1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table1_n0 +POSTHOOK: Input: default@lv_table2 +#### A masked pattern was here #### +abc 100 t test test test test test test test test test [1,2,3] 1 diff --git ql/src/test/results/clientpositive/llap/lateral_view_onview2.q.out ql/src/test/results/clientpositive/llap/lateral_view_onview2.q.out new file mode 100644 index 0000000000..2cebb14bb2 --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_onview2.q.out @@ -0,0 +1,117 @@ +PREHOOK: query: CREATE TABLE lv_table_n1( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_table_n1 +POSTHOOK: query: CREATE TABLE lv_table_n1( c1 STRING, c2 ARRAY, c3 INT, c4 CHAR(1)) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_table_n1 +PREHOOK: query: INSERT OVERWRITE TABLE lv_table_n1 SELECT 'abc ', array(1,2,3), 100, 't' FROM src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@lv_table_n1 +POSTHOOK: query: INSERT OVERWRITE TABLE lv_table_n1 SELECT 'abc ', array(1,2,3), 100, 't' FROM src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@lv_table_n1 +POSTHOOK: Lineage: lv_table_n1.c1 SIMPLE [] +POSTHOOK: Lineage: lv_table_n1.c2 EXPRESSION [] +POSTHOOK: Lineage: lv_table_n1.c3 SIMPLE [] +POSTHOOK: Lineage: lv_table_n1.c4 EXPRESSION [] +PREHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n1 +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@lv_table_n1 +PREHOOK: Output: database:default +PREHOOK: Output: default@lv_view +POSTHOOK: query: CREATE OR REPLACE VIEW lv_view AS SELECT * FROM lv_table_n1 +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@lv_table_n1 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@lv_view +POSTHOOK: Lineage: lv_view.c1 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c1, type:string, comment:null), ] +POSTHOOK: Lineage: lv_view.c2 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c2, type:array, comment:null), ] +POSTHOOK: Lineage: lv_view.c3 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c3, type:int, comment:null), ] +POSTHOOK: Lineage: lv_view.c4 SIMPLE [(lv_table_n1)lv_table_n1.FieldSchema(name:c4, type:char(1), comment:null), ] +PREHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +PREHOOK: type: QUERY +PREHOOK: Input: default@lv_table_n1 +PREHOOK: Input: default@lv_view +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT myTable.myCol, myTable2.myCol2 FROM lv_view LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array('a', 'b', 'c')) myTable2 AS myCol2 LIMIT 9 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@lv_table_n1 +POSTHOOK: Input: default@lv_view +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 9 + Processor Tree: + TableScan + alias: lv_table_n1 + Select Operator + Lateral View Forward + Select Operator + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col4 + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array('a','b','c') (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4 + Lateral View Forward + Select Operator + expressions: _col4 (type: int) + outputColumnNames: _col4 + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + Select Operator + expressions: array('a','b','c') (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col4, _col5 + Select Operator + expressions: _col4 (type: int), _col5 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 9 + ListSink + diff --git ql/src/test/results/clientpositive/llap/lateral_view_outer.q.out ql/src/test/results/clientpositive/llap/lateral_view_outer.q.out new file mode 100644 index 0000000000..31bb6f4b6c --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_outer.q.out @@ -0,0 +1,207 @@ +PREHOOK: query: explain +select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + TableScan + alias: src + Lateral View Forward + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + Select Operator + expressions: array() (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + outer lateral view: true + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + +PREHOOK: query: select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from src LATERAL VIEW OUTER explode(array()) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +238 val_238 NULL +86 val_86 NULL +311 val_311 NULL +27 val_27 NULL +165 val_165 NULL +409 val_409 NULL +255 val_255 NULL +278 val_278 NULL +98 val_98 NULL +484 val_484 NULL +PREHOOK: query: explain +select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + TableScan + alias: src + Lateral View Forward + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + Select Operator + expressions: array(4,5) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + outer lateral view: true + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + +PREHOOK: query: select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from src LATERAL VIEW OUTER explode(array(4,5)) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +238 val_238 4 +238 val_238 5 +86 val_86 4 +86 val_86 5 +311 val_311 4 +311 val_311 5 +27 val_27 4 +27 val_27 5 +165 val_165 4 +165 val_165 5 +PREHOOK: query: create table array_valued as select key, if (key > 300, array(value, value), null) as value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@array_valued +POSTHOOK: query: create table array_valued as select key, if (key > 300, array(value, value), null) as value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@array_valued +POSTHOOK: Lineage: array_valued.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: array_valued.value EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), (src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@array_valued +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@array_valued +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + TableScan + alias: array_valued + Lateral View Forward + Select Operator + expressions: key (type: string), value (type: array) + outputColumnNames: key, value + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: array), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + Select Operator + expressions: value (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + outer lateral view: true + Lateral View Join Operator + outputColumnNames: _col0, _col1, _col5 + Select Operator + expressions: _col0 (type: string), _col1 (type: array), _col5 (type: string) + outputColumnNames: _col0, _col1, _col2 + Limit + Number of rows: 10 + ListSink + +PREHOOK: query: select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@array_valued +#### A masked pattern was here #### +POSTHOOK: query: select * from array_valued LATERAL VIEW OUTER explode(value) C AS a limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@array_valued +#### A masked pattern was here #### +238 NULL NULL +86 NULL NULL +311 ["val_311","val_311"] val_311 +311 ["val_311","val_311"] val_311 +27 NULL NULL +165 NULL NULL +409 ["val_409","val_409"] val_409 +409 ["val_409","val_409"] val_409 +255 NULL NULL +278 NULL NULL diff --git ql/src/test/results/clientpositive/llap/lateral_view_ppd.q.out ql/src/test/results/clientpositive/llap/lateral_view_ppd.q.out new file mode 100644 index 0000000000..f2e128903a --- /dev/null +++ ql/src/test/results/clientpositive/llap/lateral_view_ppd.q.out @@ -0,0 +1,356 @@ +PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '0') (type: boolean) + Filter Operator + predicate: (key = '0') (type: boolean) + Lateral View Forward + Select Operator + expressions: value (type: string) + outputColumnNames: value + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +val_0 1 +val_0 2 +val_0 3 +val_0 1 +val_0 2 +val_0 3 +val_0 1 +val_0 2 +val_0 3 +PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '0') (type: boolean) + Filter Operator + predicate: (key = '0') (type: boolean) + Lateral View Forward + Select Operator + expressions: value (type: string) + outputColumnNames: value + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), 1 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Filter Operator + predicate: (col = 1) (type: boolean) + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), 1 (type: int) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE key='0' AND myCol=1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +val_0 1 +val_0 1 +val_0 1 +PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 12 + Processor Tree: + TableScan + alias: srcpart + filterExpr: ((ds = '2008-04-08') and (hr = '12')) (type: boolean) + Lateral View Forward + Select Operator + expressions: value (type: string) + outputColumnNames: value + Lateral View Join Operator + outputColumnNames: _col1, _col7 + Select Operator + expressions: _col1 (type: string), _col7 (type: int) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 12 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col1, _col7 + Select Operator + expressions: _col1 (type: string), _col7 (type: int) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 12 + ListSink + +PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM srcpart LATERAL VIEW explode(array(1,2,3)) myTable AS myCol) a WHERE ds='2008-04-08' AND hr="12" LIMIT 12 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +val_238 1 +val_238 2 +val_238 3 +val_86 1 +val_86 2 +val_86 3 +val_311 1 +val_311 2 +val_311 3 +val_27 1 +val_27 2 +val_27 3 +PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '0') (type: boolean) + Filter Operator + predicate: (key = '0') (type: boolean) + Lateral View Forward + Select Operator + expressions: value (type: string) + outputColumnNames: value + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Lateral View Forward + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col1, _col5 + Lateral View Join Operator + outputColumnNames: _col1, _col5, _col6 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col1, _col5, _col6 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Lateral View Forward + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col1, _col5 + Lateral View Join Operator + outputColumnNames: _col1, _col5, _col6 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col1, _col5, _col6 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol LATERAL VIEW explode(array(1,2,3)) myTable2 AS myCol2) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +val_0 1 +val_0 1 +val_0 1 +val_0 2 +val_0 2 +val_0 2 +val_0 3 +val_0 3 +val_0 3 +val_0 1 +val_0 1 +val_0 1 +val_0 2 +val_0 2 +val_0 2 +val_0 3 +val_0 3 +val_0 3 +val_0 1 +val_0 1 +val_0 1 +val_0 2 +val_0 2 +val_0 2 +val_0 3 +val_0 3 +val_0 3 +PREHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '0') (type: boolean) + Filter Operator + predicate: (key = '0') (type: boolean) + Lateral View Forward + Select Operator + expressions: value (type: string) + outputColumnNames: value + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + Select Operator + expressions: array(1,2,3) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Filter Operator + predicate: (col > 1) (type: boolean) + Lateral View Join Operator + outputColumnNames: _col1, _col5 + Select Operator + expressions: _col1 (type: string), _col5 (type: int) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: SELECT value, myCol FROM (SELECT * FROM src LATERAL VIEW explode(array(1,2,3)) myTable AS myCol WHERE myCol > 1) a WHERE key='0' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +val_0 2 +val_0 3 +val_0 2 +val_0 3 +val_0 2 +val_0 3 diff --git ql/src/test/results/clientpositive/leadlag_queries.q.out ql/src/test/results/clientpositive/llap/leadlag_queries.q.out similarity index 100% rename from ql/src/test/results/clientpositive/leadlag_queries.q.out rename to ql/src/test/results/clientpositive/llap/leadlag_queries.q.out index c8d23e04f3..46589e8088 100644 --- ql/src/test/results/clientpositive/leadlag_queries.q.out +++ ql/src/test/results/clientpositive/llap/leadlag_queries.q.out @@ -20,30 +20,30 @@ POSTHOOK: Input: default@part #### A masked pattern was here #### Manufacturer#1 1173.15 1173.15 1173.15 1173.15 1173.15 0.0 Manufacturer#1 1173.15 1753.76 1753.76 1753.76 1753.76 -580.6099999999999 -Manufacturer#1 1753.76 1602.59 1602.59 1602.59 1602.59 151.17000000000007 -Manufacturer#1 1602.59 1414.42 1414.42 1414.42 1414.42 188.16999999999985 Manufacturer#1 1414.42 1632.66 1632.66 1632.66 1632.66 -218.24 +Manufacturer#1 1602.59 1414.42 1414.42 1414.42 1414.42 188.16999999999985 Manufacturer#1 1632.66 NULL NULL 10.0 1632.66 0.0 +Manufacturer#1 1753.76 1602.59 1602.59 1602.59 1602.59 151.17000000000007 Manufacturer#2 1690.68 1800.7 1800.7 1800.7 1800.7 -110.01999999999998 -Manufacturer#2 1800.7 2031.98 2031.98 2031.98 2031.98 -231.27999999999997 -Manufacturer#2 2031.98 1698.66 1698.66 1698.66 1698.66 333.31999999999994 Manufacturer#2 1698.66 1701.6 1701.6 1701.6 1701.6 -2.939999999999827 Manufacturer#2 1701.6 NULL NULL 10.0 1701.6 0.0 -Manufacturer#3 1671.68 1190.27 1190.27 1190.27 1190.27 481.4100000000001 +Manufacturer#2 1800.7 2031.98 2031.98 2031.98 2031.98 -231.27999999999997 +Manufacturer#2 2031.98 1698.66 1698.66 1698.66 1698.66 333.31999999999994 Manufacturer#3 1190.27 1410.39 1410.39 1410.39 1410.39 -220.12000000000012 +Manufacturer#3 1337.29 NULL NULL 10.0 1337.29 0.0 Manufacturer#3 1410.39 1922.98 1922.98 1922.98 1922.98 -512.5899999999999 +Manufacturer#3 1671.68 1190.27 1190.27 1190.27 1190.27 481.4100000000001 Manufacturer#3 1922.98 1337.29 1337.29 1337.29 1337.29 585.69 -Manufacturer#3 1337.29 NULL NULL 10.0 1337.29 0.0 -Manufacturer#4 1620.67 1375.42 1375.42 1375.42 1375.42 245.25 -Manufacturer#4 1375.42 1206.26 1206.26 1206.26 1206.26 169.16000000000008 Manufacturer#4 1206.26 1844.92 1844.92 1844.92 1844.92 -638.6600000000001 -Manufacturer#4 1844.92 1290.35 1290.35 1290.35 1290.35 554.5700000000002 Manufacturer#4 1290.35 NULL NULL 10.0 1290.35 0.0 -Manufacturer#5 1789.69 1611.66 1611.66 1611.66 1611.66 178.02999999999997 -Manufacturer#5 1611.66 1788.73 1788.73 1788.73 1788.73 -177.06999999999994 -Manufacturer#5 1788.73 1018.1 1018.1 1018.1 1018.1 770.63 +Manufacturer#4 1375.42 1206.26 1206.26 1206.26 1206.26 169.16000000000008 +Manufacturer#4 1620.67 1375.42 1375.42 1375.42 1375.42 245.25 +Manufacturer#4 1844.92 1290.35 1290.35 1290.35 1290.35 554.5700000000002 Manufacturer#5 1018.1 1464.48 1464.48 1464.48 1464.48 -446.38 Manufacturer#5 1464.48 NULL NULL 10.0 1464.48 0.0 +Manufacturer#5 1611.66 1788.73 1788.73 1788.73 1788.73 -177.06999999999994 +Manufacturer#5 1788.73 1018.1 1018.1 1018.1 1018.1 770.63 +Manufacturer#5 1789.69 1611.66 1611.66 1611.66 1611.66 178.02999999999997 PREHOOK: query: select p_mfgr, p_name, p_retailprice, lead(p_retailprice,1) over (partition by p_mfgr, p_name ), p_retailprice - lead(p_retailprice,1,p_retailprice) over (partition by p_mfgr, p_name) @@ -104,32 +104,32 @@ from part POSTHOOK: type: QUERY POSTHOOK: Input: default@part #### A masked pattern was here #### -Manufacturer#1 1173.15 NULL NULL 1173.15 10.0 0.0 Manufacturer#1 1173.15 1173.15 1173.15 1173.15 1173.15 0.0 -Manufacturer#1 1753.76 1173.15 1173.15 1173.15 1173.15 580.6099999999999 -Manufacturer#1 1602.59 1753.76 1753.76 1753.76 1753.76 -151.17000000000007 +Manufacturer#1 1173.15 NULL NULL 1173.15 10.0 0.0 Manufacturer#1 1414.42 1602.59 1602.59 1602.59 1602.59 -188.16999999999985 +Manufacturer#1 1602.59 1753.76 1753.76 1753.76 1753.76 -151.17000000000007 Manufacturer#1 1632.66 1414.42 1414.42 1414.42 1414.42 218.24 +Manufacturer#1 1753.76 1173.15 1173.15 1173.15 1173.15 580.6099999999999 Manufacturer#2 1690.68 NULL NULL 1690.68 10.0 0.0 -Manufacturer#2 1800.7 1690.68 1690.68 1690.68 1690.68 110.01999999999998 -Manufacturer#2 2031.98 1800.7 1800.7 1800.7 1800.7 231.27999999999997 Manufacturer#2 1698.66 2031.98 2031.98 2031.98 2031.98 -333.31999999999994 Manufacturer#2 1701.6 1698.66 1698.66 1698.66 1698.66 2.939999999999827 -Manufacturer#3 1671.68 NULL NULL 1671.68 10.0 0.0 +Manufacturer#2 1800.7 1690.68 1690.68 1690.68 1690.68 110.01999999999998 +Manufacturer#2 2031.98 1800.7 1800.7 1800.7 1800.7 231.27999999999997 Manufacturer#3 1190.27 1671.68 1671.68 1671.68 1671.68 -481.4100000000001 +Manufacturer#3 1337.29 1922.98 1922.98 1922.98 1922.98 -585.69 Manufacturer#3 1410.39 1190.27 1190.27 1190.27 1190.27 220.12000000000012 +Manufacturer#3 1671.68 NULL NULL 1671.68 10.0 0.0 Manufacturer#3 1922.98 1410.39 1410.39 1410.39 1410.39 512.5899999999999 -Manufacturer#3 1337.29 1922.98 1922.98 1922.98 1922.98 -585.69 -Manufacturer#4 1620.67 NULL NULL 1620.67 10.0 0.0 -Manufacturer#4 1375.42 1620.67 1620.67 1620.67 1620.67 -245.25 Manufacturer#4 1206.26 1375.42 1375.42 1375.42 1375.42 -169.16000000000008 -Manufacturer#4 1844.92 1206.26 1206.26 1206.26 1206.26 638.6600000000001 Manufacturer#4 1290.35 1844.92 1844.92 1844.92 1844.92 -554.5700000000002 -Manufacturer#5 1789.69 NULL NULL 1789.69 10.0 0.0 -Manufacturer#5 1611.66 1789.69 1789.69 1789.69 1789.69 -178.02999999999997 -Manufacturer#5 1788.73 1611.66 1611.66 1611.66 1611.66 177.06999999999994 +Manufacturer#4 1375.42 1620.67 1620.67 1620.67 1620.67 -245.25 +Manufacturer#4 1620.67 NULL NULL 1620.67 10.0 0.0 +Manufacturer#4 1844.92 1206.26 1206.26 1206.26 1206.26 638.6600000000001 Manufacturer#5 1018.1 1788.73 1788.73 1788.73 1788.73 -770.63 Manufacturer#5 1464.48 1018.1 1018.1 1018.1 1018.1 446.38 +Manufacturer#5 1611.66 1789.69 1789.69 1789.69 1789.69 -178.02999999999997 +Manufacturer#5 1788.73 1611.66 1611.66 1611.66 1611.66 177.06999999999994 +Manufacturer#5 1789.69 NULL NULL 1789.69 10.0 0.0 PREHOOK: query: select p_mfgr, p_name, p_retailprice, lag(p_retailprice,1) over (partition by p_mfgr, p_name ), p_retailprice - lag(p_retailprice,1,p_retailprice) over (partition by p_mfgr, p_name) @@ -144,8 +144,8 @@ from part POSTHOOK: type: QUERY POSTHOOK: Input: default@part #### A masked pattern was here #### -Manufacturer#1 almond antique burnished rose metallic 1173.15 NULL 0.0 Manufacturer#1 almond antique burnished rose metallic 1173.15 1173.15 0.0 +Manufacturer#1 almond antique burnished rose metallic 1173.15 NULL 0.0 Manufacturer#1 almond antique chartreuse lavender yellow 1753.76 NULL 0.0 Manufacturer#1 almond antique salmon chartreuse burlywood 1602.59 NULL 0.0 Manufacturer#1 almond aquamarine burnished black steel 1414.42 NULL 0.0 @@ -192,27 +192,27 @@ POSTHOOK: Input: default@part #### A masked pattern was here #### Manufacturer#1 1173.15 1173.15 1173.15 0.0 NULL 1173.15 Manufacturer#1 1173.15 1753.76 1753.76 -580.6099999999999 1173.15 1173.15 -Manufacturer#1 1753.76 1602.59 1602.59 151.17000000000007 1173.15 1173.15 -Manufacturer#1 1602.59 1414.42 1414.42 188.16999999999985 1753.76 1753.76 Manufacturer#1 1414.42 1632.66 1632.66 -218.24 1602.59 1602.59 +Manufacturer#1 1602.59 1414.42 1414.42 188.16999999999985 1753.76 1753.76 Manufacturer#1 1632.66 NULL 1632.66 0.0 1414.42 1414.42 +Manufacturer#1 1753.76 1602.59 1602.59 151.17000000000007 1173.15 1173.15 Manufacturer#2 1690.68 1800.7 1800.7 -110.01999999999998 NULL 1690.68 -Manufacturer#2 1800.7 2031.98 2031.98 -231.27999999999997 1690.68 1690.68 -Manufacturer#2 2031.98 1698.66 1698.66 333.31999999999994 1800.7 1800.7 Manufacturer#2 1698.66 1701.6 1701.6 -2.939999999999827 2031.98 2031.98 Manufacturer#2 1701.6 NULL 1701.6 0.0 1698.66 1698.66 -Manufacturer#3 1671.68 1190.27 1190.27 481.4100000000001 NULL 1671.68 +Manufacturer#2 1800.7 2031.98 2031.98 -231.27999999999997 1690.68 1690.68 +Manufacturer#2 2031.98 1698.66 1698.66 333.31999999999994 1800.7 1800.7 Manufacturer#3 1190.27 1410.39 1410.39 -220.12000000000012 1671.68 1671.68 +Manufacturer#3 1337.29 NULL 1337.29 0.0 1922.98 1922.98 Manufacturer#3 1410.39 1922.98 1922.98 -512.5899999999999 1190.27 1190.27 +Manufacturer#3 1671.68 1190.27 1190.27 481.4100000000001 NULL 1671.68 Manufacturer#3 1922.98 1337.29 1337.29 585.69 1410.39 1410.39 -Manufacturer#3 1337.29 NULL 1337.29 0.0 1922.98 1922.98 -Manufacturer#4 1620.67 1375.42 1375.42 245.25 NULL 1620.67 -Manufacturer#4 1375.42 1206.26 1206.26 169.16000000000008 1620.67 1620.67 Manufacturer#4 1206.26 1844.92 1844.92 -638.6600000000001 1375.42 1375.42 -Manufacturer#4 1844.92 1290.35 1290.35 554.5700000000002 1206.26 1206.26 Manufacturer#4 1290.35 NULL 1290.35 0.0 1844.92 1844.92 -Manufacturer#5 1789.69 1611.66 1611.66 178.02999999999997 NULL 1789.69 -Manufacturer#5 1611.66 1788.73 1788.73 -177.06999999999994 1789.69 1789.69 -Manufacturer#5 1788.73 1018.1 1018.1 770.63 1611.66 1611.66 +Manufacturer#4 1375.42 1206.26 1206.26 169.16000000000008 1620.67 1620.67 +Manufacturer#4 1620.67 1375.42 1375.42 245.25 NULL 1620.67 +Manufacturer#4 1844.92 1290.35 1290.35 554.5700000000002 1206.26 1206.26 Manufacturer#5 1018.1 1464.48 1464.48 -446.38 1788.73 1788.73 Manufacturer#5 1464.48 NULL 1464.48 0.0 1018.1 1018.1 +Manufacturer#5 1611.66 1788.73 1788.73 -177.06999999999994 1789.69 1789.69 +Manufacturer#5 1788.73 1018.1 1018.1 770.63 1611.66 1611.66 +Manufacturer#5 1789.69 1611.66 1611.66 178.02999999999997 NULL 1789.69 diff --git ql/src/test/results/clientpositive/llap/limit_pushdown2.q.out ql/src/test/results/clientpositive/llap/limit_pushdown2.q.out new file mode 100644 index 0000000000..f1636bdc5d --- /dev/null +++ ql/src/test/results/clientpositive/llap/limit_pushdown2.q.out @@ -0,0 +1,1363 @@ +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 1.0 +10 val_10 11.0 +100 val_100 101.0 +103 val_103 104.0 +104 val_104 105.0 +105 val_105 106.0 +11 val_11 12.0 +111 val_111 112.0 +113 val_113 114.0 +114 val_114 115.0 +116 val_116 117.0 +118 val_118 119.0 +119 val_119 120.0 +12 val_12 13.0 +120 val_120 121.0 +125 val_125 126.0 +126 val_126 127.0 +128 val_128 129.0 +129 val_129 130.0 +131 val_131 132.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key, value desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key, value desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: +- + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: +- + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key, value desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key, value desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 1.0 +10 val_10 11.0 +100 val_100 101.0 +103 val_103 104.0 +104 val_104 105.0 +105 val_105 106.0 +11 val_11 12.0 +111 val_111 112.0 +113 val_113 114.0 +114 val_114 115.0 +116 val_116 117.0 +118 val_118 119.0 +119 val_119 120.0 +12 val_12 13.0 +120 val_120 121.0 +125 val_125 126.0 +126 val_126 127.0 +128 val_128 129.0 +129 val_129 130.0 +131 val_131 132.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key desc, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by key, value +order by key desc, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: -+ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key desc, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by key, value +order by key desc, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +98 val_98 99.0 +97 val_97 98.0 +96 val_96 97.0 +95 val_95 96.0 +92 val_92 93.0 +90 val_90 91.0 +9 val_9 10.0 +87 val_87 88.0 +86 val_86 87.0 +85 val_85 86.0 +84 val_84 85.0 +83 val_83 84.0 +82 val_82 83.0 +80 val_80 81.0 +8 val_8 9.0 +78 val_78 79.0 +77 val_77 78.0 +76 val_76 77.0 +74 val_74 75.0 +72 val_72 73.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col1 (type: string), _col0 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 1.0 +10 val_10 11.0 +100 val_100 101.0 +103 val_103 104.0 +104 val_104 105.0 +105 val_105 106.0 +11 val_11 12.0 +111 val_111 112.0 +113 val_113 114.0 +114 val_114 115.0 +116 val_116 117.0 +118 val_118 119.0 +119 val_119 120.0 +12 val_12 13.0 +120 val_120 121.0 +125 val_125 126.0 +126 val_126 127.0 +128 val_128 129.0 +129 val_129 130.0 +131 val_131 132.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key desc, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key desc, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: -+ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col1 (type: string), _col0 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key desc, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key desc, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +98 val_98 99.0 +97 val_97 98.0 +96 val_96 97.0 +95 val_95 96.0 +92 val_92 93.0 +90 val_90 91.0 +9 val_9 10.0 +87 val_87 88.0 +86 val_86 87.0 +85 val_85 86.0 +84 val_84 85.0 +83 val_83 84.0 +82 val_82 83.0 +80 val_80 81.0 +8 val_8 9.0 +78 val_78 79.0 +77 val_77 78.0 +76 val_76 77.0 +74 val_74 75.0 +72 val_72 73.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key +order by key desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: -+ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col1 (type: string), _col0 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double), _col3 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 250 Data size: 48500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), (_col2 / _col3) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by value, key +order by key desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +98 val_98 99.0 +97 val_97 98.0 +96 val_96 97.0 +95 val_95 96.0 +92 val_92 93.0 +90 val_90 91.0 +9 val_9 10.0 +87 val_87 88.0 +86 val_86 87.0 +85 val_85 86.0 +84 val_84 85.0 +83 val_83 84.0 +82 val_82 83.0 +80 val_80 81.0 +8 val_8 9.0 +78 val_78 79.0 +77 val_77 78.0 +76 val_76 77.0 +74 val_74 75.0 +72 val_72 73.0 +PREHOOK: query: explain +select key, value, count(key + 1) as agg1 from src +group by key, value +order by key, value, agg1 limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, count(key + 1) as agg1 from src +group by key, value +order by key, value, agg1 limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: ++ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: +++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, count(key + 1) as agg1 from src +group by key, value +order by key, value, agg1 limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, count(key + 1) as agg1 from src +group by key, value +order by key, value, agg1 limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 3 +10 val_10 1 +100 val_100 2 +103 val_103 2 +104 val_104 2 +105 val_105 1 +11 val_11 1 +111 val_111 1 +113 val_113 2 +114 val_114 1 +116 val_116 1 +118 val_118 2 +119 val_119 3 +12 val_12 2 +120 val_120 2 +125 val_125 2 +126 val_126 1 +128 val_128 3 +129 val_129 2 +131 val_131 1 +PREHOOK: query: explain +select key, value, count(key + 1) as agg1 from src +group by key, value +order by key desc, value, agg1 limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, count(key + 1) as agg1 from src +group by key, value +order by key desc, value, agg1 limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: -+ + keys: key (type: string), value (type: string) + null sort order: zz + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: count(_col2) + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: -++ + keys: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: -++ + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), KEY.reducesinkkey2 (type: bigint) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 46500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, count(key + 1) as agg1 from src +group by key, value +order by key desc, value, agg1 limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, count(key + 1) as agg1 from src +group by key, value +order by key desc, value, agg1 limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +98 val_98 2 +97 val_97 2 +96 val_96 1 +95 val_95 2 +92 val_92 1 +90 val_90 3 +9 val_9 1 +87 val_87 1 +86 val_86 1 +85 val_85 1 +84 val_84 2 +83 val_83 2 +82 val_82 1 +80 val_80 1 +8 val_8 1 +78 val_78 1 +77 val_77 1 +76 val_76 2 +74 val_74 1 +72 val_72 2 +PREHOOK: query: explain +select value, avg(key + 1) myavg from src +group by value +order by myavg, value desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select value, avg(key + 1) myavg from src +group by value +order by myavg, value desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1), count(_col1) + keys: _col0 (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: double), _col2 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + Top N Key Operator + sort order: +- + keys: (_col1 / _col2) (type: double), _col0 (type: string) + null sort order: zz + Statistics: Num rows: 250 Data size: 26750 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: _col0 (type: string), (_col1 / _col2) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double), _col0 (type: string) + null sort order: zz + sort order: +- + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey1 (type: string), KEY.reducesinkkey0 (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select value, avg(key + 1) myavg from src +group by value +order by myavg, value desc limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select value, avg(key + 1) myavg from src +group by value +order by myavg, value desc limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +val_0 1.0 +val_2 3.0 +val_4 5.0 +val_5 6.0 +val_8 9.0 +val_9 10.0 +val_10 11.0 +val_11 12.0 +val_12 13.0 +val_15 16.0 +val_17 18.0 +val_18 19.0 +val_19 20.0 +val_20 21.0 +val_24 25.0 +val_26 27.0 +val_27 28.0 +val_28 29.0 +val_30 31.0 +val_33 34.0 +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key with rollup +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by value, key with rollup +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: double), _col4 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3, _col4 + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Top N Key Operator + sort order: ++ + keys: _col1 (type: string), _col0 (type: string) + null sort order: zz + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: _col1 (type: string), _col0 (type: string), (_col3 / _col4) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: explain +select key, value, avg(key + 1) from src +group by rollup(value, key) +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +select key, value, avg(key + 1) from src +group by rollup(value, key) +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: value (type: string), key (type: string), (UDFToDouble(key) + 1.0D) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col2), count(_col2) + keys: _col0 (type: string), _col1 (type: string), 0L (type: bigint) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + null sort order: zzz + sort order: +++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: bigint) + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col3 (type: double), _col4 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0), count(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: bigint) + mode: mergepartial + outputColumnNames: _col0, _col1, _col3, _col4 + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + pruneGroupingSetId: true + Top N Key Operator + sort order: ++ + keys: _col1 (type: string), _col0 (type: string) + null sort order: zz + Statistics: Num rows: 750 Data size: 151500 Basic stats: COMPLETE Column stats: COMPLETE + top n: 20 + Select Operator + expressions: _col1 (type: string), _col0 (type: string), (_col3 / _col4) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col2 (type: double) + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: string), KEY.reducesinkkey1 (type: string), VALUE._col0 (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 750 Data size: 139500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 3720 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select key, value, avg(key + 1) from src +group by value, key with rollup +order by key, value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value, avg(key + 1) from src +group by value, key with rollup +order by key, value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 val_0 1.0 +10 val_10 11.0 +100 val_100 101.0 +103 val_103 104.0 +104 val_104 105.0 +105 val_105 106.0 +11 val_11 12.0 +111 val_111 112.0 +113 val_113 114.0 +114 val_114 115.0 +116 val_116 117.0 +118 val_118 119.0 +119 val_119 120.0 +12 val_12 13.0 +120 val_120 121.0 +125 val_125 126.0 +126 val_126 127.0 +128 val_128 129.0 +129 val_129 130.0 +131 val_131 132.0 diff --git ql/src/test/results/clientpositive/llap/limit_pushdown_negative.q.out ql/src/test/results/clientpositive/llap/limit_pushdown_negative.q.out new file mode 100644 index 0000000000..49b7a6b2df --- /dev/null +++ ql/src/test/results/clientpositive/llap/limit_pushdown_negative.q.out @@ -0,0 +1,440 @@ +PREHOOK: query: explain select * from src a join src b on a.key=b.key limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from src a join src b on a.key=b.key limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 791 Data size: 281596 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 7120 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 7120 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: explain select value, sum(key) as sum from src group by value having sum > 100 limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select value, sum(key) as sum from src group by value having sum > 100 limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(key) + keys: value (type: string) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: double) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 24750 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (_col1 > 100.0D) (type: boolean) + Statistics: Num rows: 83 Data size: 8217 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: explain select key, L.* from (select * from src order by key) a lateral view explode(array(value, value)) L as v limit 10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select key, L.* from (select * from src order by key) a lateral view explode(array(value, value)) L as v limit 10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 10 + Processor Tree: + TableScan + alias: src + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Lateral View Forward + Select Operator + expressions: _col0 (type: string) + outputColumnNames: _col0 + Lateral View Join Operator + outputColumnNames: _col0, _col2 + Select Operator + expressions: _col0 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 10 + ListSink + Select Operator + expressions: array(_col1,_col1) (type: array) + outputColumnNames: _col0 + UDTF Operator + function name: explode + Lateral View Join Operator + outputColumnNames: _col0, _col2 + Select Operator + expressions: _col0 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1 + Limit + Number of rows: 10 + ListSink + +PREHOOK: query: CREATE TABLE dest_2(key STRING, c1 INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_2 +POSTHOOK: query: CREATE TABLE dest_2(key STRING, c1 INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_2 +PREHOOK: query: CREATE TABLE dest_3(key STRING, c1 INT) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_3 +POSTHOOK: query: CREATE TABLE dest_3(key STRING, c1 INT) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_3 +PREHOOK: query: EXPLAIN FROM src +INSERT OVERWRITE TABLE dest_2 SELECT value, sum(key) GROUP BY value +INSERT OVERWRITE TABLE dest_3 SELECT value, sum(key) GROUP BY value limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@dest_2 +PREHOOK: Output: default@dest_3 +POSTHOOK: query: EXPLAIN FROM src +INSERT OVERWRITE TABLE dest_2 SELECT value, sum(key) GROUP BY value +INSERT OVERWRITE TABLE dest_3 SELECT value, sum(key) GROUP BY value limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@dest_2 +POSTHOOK: Output: default@dest_3 +STAGE DEPENDENCIES: + Stage-2 is a root stage + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-2 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) + Reducer 5 <- Reducer 4 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: value (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: value (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Forward + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_2 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, c1 + Statistics: Num rows: 307 Data size: 29165 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string) + mode: complete + outputColumnNames: _col0, _col1 + Statistics: Num rows: 307 Data size: 30393 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.3 + value expressions: _col0 (type: string), _col1 (type: double) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 1980 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), UDFToInteger(_col1) (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_3 + Select Operator + expressions: _col0 (type: string), _col1 (type: int) + outputColumnNames: key, c1 + Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(c1, 'hll') + minReductionHashAggr: 0.95 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-3 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_2 + + Stage: Stage-4 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, c1 + Column Types: string, int + Table: default.dest_2 + + Stage: Stage-1 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_3 + + Stage: Stage-5 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, c1 + Column Types: string, int + Table: default.dest_3 + diff --git ql/src/test/results/clientpositive/llap/lineage1.q.out ql/src/test/results/clientpositive/llap/lineage1.q.out new file mode 100644 index 0000000000..cff56bb936 --- /dev/null +++ ql/src/test/results/clientpositive/llap/lineage1.q.out @@ -0,0 +1,289 @@ +PREHOOK: query: CREATE TABLE dest_l1(key INT, value STRING) STORED AS TEXTFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@dest_l1 +POSTHOOK: query: CREATE TABLE dest_l1(key INT, value STRING) STORED AS TEXTFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@dest_l1 +PREHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_l1 +SELECT j.* +FROM (SELECT t1.key, p1.value + FROM src1 t1 + LEFT OUTER JOIN src p1 + ON (t1.key = p1.key) + UNION ALL + SELECT t2.key, p2.value + FROM src1 t2 + LEFT OUTER JOIN src p2 + ON (t2.key = p2.key)) j +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_l1 +POSTHOOK: query: EXPLAIN +INSERT OVERWRITE TABLE dest_l1 +SELECT j.* +FROM (SELECT t1.key, p1.value + FROM src1 t1 + LEFT OUTER JOIN src p1 + ON (t1.key = p1.key) + UNION ALL + SELECT t2.key, p2.value + FROM src1 t2 + LEFT OUTER JOIN src p2 + ON (t2.key = p2.key)) j +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_l1 +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 5 (SIMPLE_EDGE), Union 3 (CONTAINS) + Reducer 4 <- Union 3 (CUSTOM_SIMPLE_EDGE) + Reducer 7 <- Map 6 (SIMPLE_EDGE), Map 8 (SIMPLE_EDGE), Union 3 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: p1 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: t1 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: p2 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Map 8 + Map Operator Tree: + TableScan + alias: t2 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 25 Data size: 2150 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_l1 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 7 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Right Outer Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col1, _col2 + Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 64 Data size: 9144 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(_col0) (type: int), _col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_l1 + Select Operator + expressions: _col0 (type: int), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 128 Data size: 5426 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 864 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct) + Union 3 + Vertex: Union 3 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest_l1 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: int, string + Table: default.dest_l1 + +PREHOOK: query: INSERT OVERWRITE TABLE dest_l1 +SELECT j.* +FROM (SELECT t1.key, p1.value + FROM src1 t1 + LEFT OUTER JOIN src p1 + ON (t1.key = p1.key) + UNION ALL + SELECT t2.key, p2.value + FROM src1 t2 + LEFT OUTER JOIN src p2 + ON (t2.key = p2.key)) j +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@src1 +PREHOOK: Output: default@dest_l1 +POSTHOOK: query: INSERT OVERWRITE TABLE dest_l1 +SELECT j.* +FROM (SELECT t1.key, p1.value + FROM src1 t1 + LEFT OUTER JOIN src p1 + ON (t1.key = p1.key) + UNION ALL + SELECT t2.key, p2.value + FROM src1 t2 + LEFT OUTER JOIN src p2 + ON (t2.key = p2.key)) j +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@src1 +POSTHOOK: Output: default@dest_l1 +POSTHOOK: Lineage: dest_l1.key EXPRESSION [(src1)t1.FieldSchema(name:key, type:string, comment:default), (src1)t2.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: dest_l1.value EXPRESSION [(src)p1.FieldSchema(name:value, type:string, comment:default), (src)p2.FieldSchema(name:value, type:string, comment:default), ] diff --git ql/src/test/results/clientpositive/list_bucket_dml_1.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out similarity index 59% rename from ql/src/test/results/clientpositive/list_bucket_dml_1.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out index 57546be38a..189335a3ad 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_1.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out @@ -30,32 +30,117 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.list_bucketing_dynamic_part_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -63,183 +148,112 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n0 + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing_dynamic_part_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -270,7 +284,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n0 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -428,57 +442,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, `value` FROM `default`.`list_bucketing_dynamic_part_n0` WHERE `key` = '484' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_dynamic_part_n0 - filterExpr: ((key = '484') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '484') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: key=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -525,14 +496,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n0 name: default.list_bucketing_dynamic_part_n0 - Truncated Path -> Alias: - /list_bucketing_dynamic_part_n0/ds=2008-04-08/hr=11/key=484 [list_bucketing_dynamic_part_n0] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_dynamic_part_n0 + filterExpr: ((key = '484') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '484') (type: boolean) + Select Operator + expressions: '484' (type: string), value (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select key, value from list_bucketing_dynamic_part_n0 where ds='2008-04-08' and hr='11' and key = "484" PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_dml_11.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out similarity index 56% rename from ql/src/test/results/clientpositive/list_bucket_dml_11.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out index 79a47dfaa6..3cfe511333 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_11.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out @@ -30,164 +30,178 @@ OPTIMIZED SQL: SELECT `key`, `value` FROM `default`.`src` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n3 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n3 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 500 Data size: 179000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_static_part_n3 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n3 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 500 Data size: 179000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -218,7 +232,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n3 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -306,57 +320,14 @@ OPTIMIZED SQL: SELECT `key`, CAST('val_466' AS STRING) AS `value` FROM `default`.`list_bucketing_static_part_n3` WHERE `value` = 'val_466' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_static_part_n3 - filterExpr: ((value = 'val_466') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (value = 'val_466') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), 'val_466' (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_466 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -403,14 +374,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n3 name: default.list_bucketing_static_part_n3 - Truncated Path -> Alias: - /list_bucketing_static_part_n3/ds=2008-04-08/hr=11/value=val_466 [list_bucketing_static_part_n3] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_static_part_n3 + filterExpr: ((value = 'val_466') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (value = 'val_466') (type: boolean) + Select Operator + expressions: key (type: string), 'val_466' (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select key, value from list_bucketing_static_part_n3 where ds='2008-04-08' and hr='11' and value = "val_466" PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_dml_12.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out similarity index 55% rename from ql/src/test/results/clientpositive/list_bucket_dml_12.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out index 012cad2116..297dd32b35 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_12.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out @@ -30,164 +30,178 @@ OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, 1 AS `_o__c2`, `value`, 1 AS `_o__c4 FROM `default`.`src` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: '1' (type: string), key (type: string), '1' (type: string), value (type: string), '1' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: '1' (type: string), key (type: string), '1' (type: string), value (type: string), '1' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns col1,col2,col3,col4,col5 + columns.comments + columns.types string:string:string:string:string +#### A masked pattern was here #### + name default.list_bucketing_mul_col_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_mul_col_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: col1, col2, col3, col4, col5, ds, hr + Statistics: Num rows: 500 Data size: 306500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll'), compute_stats(col4, 'hll'), compute_stats(col5, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , - columns col1,col2,col3,col4,col5 - columns.comments - columns.types string:string:string:string:string -#### A masked pattern was here #### - name default.list_bucketing_mul_col_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_mul_col_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: col1, col2, col3, col4, col5, ds, hr - Statistics: Num rows: 500 Data size: 306500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll'), compute_stats(col4, 'hll'), compute_stats(col5, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2380 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types struct:struct:struct:struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 + columns.types struct:struct:struct:struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -218,7 +232,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -314,57 +328,14 @@ OPTIMIZED SQL: SELECT `col1`, CAST('466' AS STRING) AS `col2`, `col3`, CAST('val FROM `default`.`list_bucketing_mul_col_n0` WHERE `col2` = '466' AND `col4` = 'val_466' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_mul_col_n0 - filterExpr: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) - Statistics: Num rows: 1 Data size: 433 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: col1 (type: string), '466' (type: string), col3 (type: string), 'val_466' (type: string), col5 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 613 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 613 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types string:string:string:string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: col4=val_466 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -411,14 +382,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 name: default.list_bucketing_mul_col_n0 - Truncated Path -> Alias: - /list_bucketing_mul_col_n0/ds=2008-04-08/hr=11/col2=466/col4=val_466 [list_bucketing_mul_col_n0] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_mul_col_n0 + filterExpr: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) + Select Operator + expressions: col1 (type: string), '466' (type: string), col3 (type: string), 'val_466' (type: string), col5 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink PREHOOK: query: select * from list_bucketing_mul_col_n0 where ds='2008-04-08' and hr='11' and col2 = "466" and col4 = "val_466" @@ -453,57 +428,14 @@ OPTIMIZED SQL: SELECT `col1`, CAST('382' AS STRING) AS `col2`, `col3`, CAST('val FROM `default`.`list_bucketing_mul_col_n0` WHERE `col2` = '382' AND `col4` = 'val_382' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_mul_col_n0 - filterExpr: ((col2 = '382') and (col4 = 'val_382')) (type: boolean) - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((col2 = '382') and (col4 = 'val_382')) (type: boolean) - Statistics: Num rows: 1 Data size: 433 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: col1 (type: string), '382' (type: string), col3 (type: string), 'val_382' (type: string), col5 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 613 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 613 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types string:string:string:string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -550,14 +482,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 name: default.list_bucketing_mul_col_n0 - Truncated Path -> Alias: - /list_bucketing_mul_col_n0/ds=2008-04-08/hr=11/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [list_bucketing_mul_col_n0] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_mul_col_n0 + filterExpr: ((col2 = '382') and (col4 = 'val_382')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((col2 = '382') and (col4 = 'val_382')) (type: boolean) + Select Operator + expressions: col1 (type: string), '382' (type: string), col3 (type: string), 'val_382' (type: string), col5 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink PREHOOK: query: select * from list_bucketing_mul_col_n0 where ds='2008-04-08' and hr='11' and col2 = "382" and col4 = "val_382" diff --git ql/src/test/results/clientpositive/list_bucket_dml_13.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out similarity index 55% rename from ql/src/test/results/clientpositive/list_bucket_dml_13.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out index 09e0eaa291..e272f2180c 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_13.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out @@ -30,164 +30,178 @@ OPTIMIZED SQL: SELECT 1 AS `_o__c0`, `key`, 1 AS `_o__c2`, `value`, 1 AS `_o__c4 FROM `default`.`src` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: '1' (type: string), key (type: string), '1' (type: string), value (type: string), '1' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: '1' (type: string), key (type: string), '1' (type: string), value (type: string), '1' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=2013-01-23+18%3A00%3A99/ + Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns col1,col2,col3,col4,col5 + columns.comments + columns.types string:string:string:string:string +#### A masked pattern was here #### + name default.list_bucketing_mul_col + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_mul_col + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), '2008-04-08' (type: string), '2013-01-23+18:00:99' (type: string) + outputColumnNames: col1, col2, col3, col4, col5, ds, hr + Statistics: Num rows: 500 Data size: 315000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll'), compute_stats(col4, 'hll'), compute_stats(col5, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=2013-01-23+18%3A00%3A99/ - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE + Path -> Partition: #### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , - columns col1,col2,col3,col4,col5 - columns.comments - columns.types string:string:string:string:string + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_mul_col - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_mul_col - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string), _col4 (type: string), '2008-04-08' (type: string), '2013-01-23+18:00:99' (type: string) - outputColumnNames: col1, col2, col3, col4, col5, ds, hr - Statistics: Num rows: 500 Data size: 315000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll'), compute_stats(col3, 'hll'), compute_stats(col4, 'hll'), compute_stats(col5, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src - Truncated Path -> Alias: - /src [src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3), compute_stats(VALUE._col4) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col4 (type: struct), _col5 (type: struct), _col6 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 2397 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types struct:struct:struct:struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 + columns.types struct:struct:struct:struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -218,7 +232,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -314,57 +328,14 @@ OPTIMIZED SQL: SELECT `col1`, CAST('466' AS STRING) AS `col2`, `col3`, CAST('val FROM `default`.`list_bucketing_mul_col` WHERE `col2` = '466' AND `col4` = 'val_466' AND `ds` = '2008-04-08' AND `hr` = '2013-01-23+18:00:99' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_mul_col - filterExpr: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) - Statistics: Num rows: 500 Data size: 216500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) - Statistics: Num rows: 1 Data size: 433 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: col1 (type: string), '466' (type: string), col3 (type: string), 'val_466' (type: string), col5 (type: string), '2008-04-08' (type: string), '2013-01-23+18:00:99' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1 Data size: 630 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 630 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3,_col4,_col5,_col6 - columns.types string:string:string:string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: col4=val_466 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -411,14 +382,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col name: default.list_bucketing_mul_col - Truncated Path -> Alias: - /list_bucketing_mul_col/ds=2008-04-08/hr=2013-01-23+18%3A00%3A99/col2=466/col4=val_466 [list_bucketing_mul_col] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_mul_col + filterExpr: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((col2 = '466') and (col4 = 'val_466')) (type: boolean) + Select Operator + expressions: col1 (type: string), '466' (type: string), col3 (type: string), 'val_466' (type: string), col5 (type: string), '2008-04-08' (type: string), '2013-01-23+18:00:99' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6 + ListSink PREHOOK: query: select * from list_bucketing_mul_col where ds='2008-04-08' and hr='2013-01-23+18:00:99' and col2 = "466" and col4 = "val_466" diff --git ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out new file mode 100644 index 0000000000..681652d8cd --- /dev/null +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out @@ -0,0 +1,359 @@ +PREHOOK: query: create table list_bucketing (key String, value String) +skewed by (key) on ("484") +stored as DIRECTORIES +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@list_bucketing +POSTHOOK: query: create table list_bucketing (key String, value String) +skewed by (key) on ("484") +stored as DIRECTORIES +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@list_bucketing +PREHOOK: query: explain extended +insert overwrite table list_bucketing select * from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@list_bucketing +POSTHOOK: query: explain extended +insert overwrite table list_bucketing select * from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@list_bucketing +OPTIMIZED SQL: SELECT `key`, `value` +FROM `default`.`src` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct list_bucketing { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.list_bucketing + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing + numFiles 0 + numRows 0 + rawDataSize 0 + serialization.ddl struct list_bucketing { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 0 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.list_bucketing + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.list_bucketing + Is Table Level Stats: true + +PREHOOK: query: insert overwrite table list_bucketing select * from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@list_bucketing +POSTHOOK: query: insert overwrite table list_bucketing select * from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@list_bucketing +POSTHOOK: Lineage: list_bucketing.key SIMPLE [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: list_bucketing.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: desc formatted list_bucketing +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@list_bucketing +POSTHOOK: query: desc formatted list_bucketing +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@list_bucketing +# col_name data_type comment +key string +value string + +# Detailed Table Information +Database: default +#### A masked pattern was here #### +Retention: 0 +#### A masked pattern was here #### +Table Type: MANAGED_TABLE +Table Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + bucketing_version 2 + numFiles 2 + numRows 500 + rawDataSize 5312 + totalSize 5812 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +InputFormat: org.apache.hadoop.mapred.TextInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Stored As SubDirectories: Yes +Skewed Columns: [key] +Skewed Values: [[484]] +#### A masked pattern was here #### +Skewed Value to Truncated Path: {[484]=/list_bucketing/key=484} +Storage Desc Params: + serialization.format 1 +PREHOOK: query: select count(1) from src +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select count(1) from src +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +500 +PREHOOK: query: select count(1) from list_bucketing +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +POSTHOOK: query: select count(1) from list_bucketing +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +500 +PREHOOK: query: select key, value from src where key = "484" +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select key, value from src where key = "484" +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +484 val_484 +PREHOOK: query: explain extended +select key, value from list_bucketing where key = "484" +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +POSTHOOK: query: explain extended +select key, value from list_bucketing where key = "484" +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, `value` +FROM `default`.`list_bucketing` +WHERE `key` = '484' +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: list_bucketing + filterExpr: (key = '484') (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '484') (type: boolean) + Select Operator + expressions: '484' (type: string), value (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select key, value from list_bucketing where key = "484" +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +POSTHOOK: query: select key, value from list_bucketing where key = "484" +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing +#### A masked pattern was here #### +484 val_484 +PREHOOK: query: drop table list_bucketing +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@list_bucketing +PREHOOK: Output: default@list_bucketing +POSTHOOK: query: drop table list_bucketing +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@list_bucketing +POSTHOOK: Output: default@list_bucketing diff --git ql/src/test/results/clientpositive/list_bucket_dml_2.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out similarity index 64% rename from ql/src/test/results/clientpositive/list_bucket_dml_2.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out index f39cadc239..bd1df7c92d 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_2.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out @@ -35,216 +35,230 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n4 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n4 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_static_part_n4 + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_static_part_n4 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -275,7 +289,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n4 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -330,10 +344,10 @@ Table: list_bucketing_static_part_n4 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 6 + numFiles 3 numRows 1000 rawDataSize 9624 - totalSize 10898 + totalSize 10659 #### A masked pattern was here #### # Storage Information @@ -385,57 +399,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) FROM `default`.`list_bucketing_static_part_n4` WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_static_part_n4 - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 358 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_484 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -450,7 +421,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n4 - numFiles 6 + numFiles 3 numRows 1000 partition_columns ds/hr partition_columns.types string:string @@ -458,7 +429,7 @@ STAGE PLANS: serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10898 + totalSize 10659 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -482,14 +453,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n4 name: default.list_bucketing_static_part_n4 - Truncated Path -> Alias: - /list_bucketing_static_part_n4/ds=2008-04-08/hr=11/key=484/value=val_484 [list_bucketing_static_part_n4] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_static_part_n4 + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from list_bucketing_static_part_n4 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_dml_3.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out similarity index 55% rename from ql/src/test/results/clientpositive/list_bucket_dml_3.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out index 1944074b71..ad1b340a9d 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_3.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out @@ -25,32 +25,166 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n1 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.list_bucketing_static_part_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/hr=11/ - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - table: + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -58,183 +192,63 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_static_part_n1 + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing_static_part_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -265,7 +279,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_static_part_n1 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -311,7 +325,7 @@ Table: list_bucketing_static_part_n1 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 4 + numFiles 2 numRows 1000 rawDataSize 10624 totalSize 11624 @@ -377,57 +391,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, `value` FROM `default`.`list_bucketing_static_part_n1` WHERE `key` = '484' AND `ds` = '2008-04-08' AND `hr` = '11' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_static_part_n1 - filterExpr: ((key = '484') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '484') (type: boolean) - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 3 Data size: 534 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: key=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -442,7 +413,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n1 - numFiles 4 + numFiles 2 numRows 1000 partition_columns ds/hr partition_columns.types string:string @@ -474,14 +445,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_static_part_n1 name: default.list_bucketing_static_part_n1 - Truncated Path -> Alias: - /list_bucketing_static_part_n1/ds=2008-04-08/hr=11/key=484 [list_bucketing_static_part_n1] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_static_part_n1 + filterExpr: ((key = '484') and (ds = '2008-04-08') and (hr = '11')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '484') (type: boolean) + Select Operator + expressions: '484' (type: string), value (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select key, value from list_bucketing_static_part_n1 where ds='2008-04-08' and hr='11' and key = "484" PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out new file mode 100644 index 0000000000..25a29beee5 --- /dev/null +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out @@ -0,0 +1,853 @@ +PREHOOK: query: create table list_bucketing_static_part_n2 (key String, value String) + partitioned by (ds String, hr String) + skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103')) + stored as DIRECTORIES + STORED AS RCFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@list_bucketing_static_part_n2 +POSTHOOK: query: create table list_bucketing_static_part_n2 (key String, value String) + partitioned by (ds String, hr String) + skewed by (key, value) on (('484','val_484'),('51','val_14'),('103','val_103')) + stored as DIRECTORIES + STORED AS RCFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@list_bucketing_static_part_n2 +PREHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +OPTIMIZED SQL: SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008-04-08 + hr 11 + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.list_bucketing_static_part_n2 + Is Table Level Stats: false + +PREHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: show partitions list_bucketing_static_part_n2 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: query: show partitions list_bucketing_static_part_n2 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@list_bucketing_static_part_n2 +ds=2008-04-08/hr=11 +PREHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n2 +# col_name data_type comment +key string +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: list_bucketing_static_part_n2 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 3 + numRows 1000 + rawDataSize 9624 + totalSize 10659 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Stored As SubDirectories: Yes +Skewed Columns: [key, value] +Skewed Values: [[103, val_103], [484, val_484], [51, val_14]] +#### A masked pattern was here #### +Skewed Value to Truncated Path: {[103, val_103]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=103/value=val_103, [484, val_484]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=484/value=val_484} +Storage Desc Params: + serialization.format 1 +PREHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +OPTIMIZED SQL: SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008-04-08 + hr 11 + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n2 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.list_bucketing_static_part_n2 + Is Table Level Stats: false + +PREHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: query: insert overwrite table list_bucketing_static_part_n2 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: list_bucketing_static_part_n2 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: show partitions list_bucketing_static_part_n2 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: query: show partitions list_bucketing_static_part_n2 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@list_bucketing_static_part_n2 +ds=2008-04-08/hr=11 +PREHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: query: desc formatted list_bucketing_static_part_n2 partition (ds='2008-04-08', hr='11') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n2 +# col_name data_type comment +key string +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: list_bucketing_static_part_n2 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 3 + numRows 1000 + rawDataSize 9624 + totalSize 10659 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Stored As SubDirectories: Yes +Skewed Columns: [key, value] +Skewed Values: [[103, val_103], [484, val_484], [51, val_14]] +#### A masked pattern was here #### +Skewed Value to Truncated Path: {[103, val_103]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=103/value=val_103, [484, val_484]=/list_bucketing_static_part_n2/ds=2008-04-08/hr=11/key=484/value=val_484} +Storage Desc Params: + serialization.format 1 +PREHOOK: query: select count(1) from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +#### A masked pattern was here #### +POSTHOOK: query: select count(1) from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +#### A masked pattern was here #### +1000 +PREHOOK: query: select count(*) from list_bucketing_static_part_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n2 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from list_bucketing_static_part_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n2 +#### A masked pattern was here #### +1000 +PREHOOK: query: explain extended +select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n2 +PREHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: explain extended +select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) AS `$f1`, CAST('2008-04-08' AS STRING) AS `$f2`, CAST('11' AS STRING) AS `$f3` +FROM `default`.`list_bucketing_static_part_n2` +WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '2008-04-08' AND `hr` = '11' +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: + Partition + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + numFiles 3 + numRows 1000 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 9624 + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + totalSize 10659 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n2 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n2 + name: default.list_bucketing_static_part_n2 + Processor Tree: + TableScan + alias: list_bucketing_static_part_n2 + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink + +PREHOOK: query: select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n2 +PREHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: select * from list_bucketing_static_part_n2 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: Input: default@list_bucketing_static_part_n2@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +484 val_484 2008-04-08 11 +484 val_484 2008-04-08 11 +PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +484 val_484 2008-04-08 11 +484 val_484 2008-04-08 12 +PREHOOK: query: drop table list_bucketing_static_part_n2 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@list_bucketing_static_part_n2 +PREHOOK: Output: default@list_bucketing_static_part_n2 +POSTHOOK: query: drop table list_bucketing_static_part_n2 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n2 +POSTHOOK: Output: default@list_bucketing_static_part_n2 diff --git ql/src/test/results/clientpositive/list_bucket_dml_5.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out similarity index 63% rename from ql/src/test/results/clientpositive/list_bucket_dml_5.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out index 11c57fd884..7aa4b23f1f 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_5.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out @@ -30,32 +30,117 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part_n1 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.list_bucketing_dynamic_part_n1 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -63,183 +148,112 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n1 + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.list_bucketing_dynamic_part_n1 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -270,7 +284,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -435,57 +449,14 @@ OPTIMIZED SQL: SELECT CAST('103' AS STRING) AS `key`, CAST('val_103' AS STRING) FROM `default`.`list_bucketing_dynamic_part_n1` WHERE `key` = '103' AND `value` = 'val_103' AND `ds` = '2008-04-08' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_dynamic_part_n1 - filterExpr: ((key = '103') and (value = 'val_103')) (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '103') and (value = 'val_103')) (type: boolean) - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '103' (type: string), 'val_103' (type: string), '2008-04-08' (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_103 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -532,9 +503,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 name: default.list_bucketing_dynamic_part_n1 -#### A masked pattern was here #### Partition - base file name: value=val_103 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -581,15 +550,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 name: default.list_bucketing_dynamic_part_n1 - Truncated Path -> Alias: - /list_bucketing_dynamic_part_n1/ds=2008-04-08/hr=11/key=103/value=val_103 [list_bucketing_dynamic_part_n1] - /list_bucketing_dynamic_part_n1/ds=2008-04-08/hr=12/key=103/value=val_103 [list_bucketing_dynamic_part_n1] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_dynamic_part_n1 + filterExpr: ((key = '103') and (value = 'val_103')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '103') and (value = 'val_103')) (type: boolean) + Select Operator + expressions: '103' (type: string), 'val_103' (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select key, value, ds, hr from list_bucketing_dynamic_part_n1 where ds='2008-04-08' and key = "103" and value ="val_103" PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_dml_6.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out similarity index 53% rename from ql/src/test/results/clientpositive/list_bucket_dml_6.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out index f7d7f2cf64..a3dccdfe80 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_6.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out @@ -34,216 +34,230 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part_n3 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_dynamic_part_n3 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part_n3 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -274,7 +288,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -333,10 +347,10 @@ Table: list_bucketing_dynamic_part_n3 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 2 + numFiles 1 numRows 16 rawDataSize 136 - totalSize 310 + totalSize 235 #### A masked pattern was here #### # Storage Information @@ -374,10 +388,10 @@ Table: list_bucketing_dynamic_part_n3 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 6 + numFiles 3 numRows 984 rawDataSize 9488 - totalSize 10734 + totalSize 10495 #### A masked pattern was here #### # Storage Information @@ -415,230 +429,230 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part_n3 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_dynamic_part_n3 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part_n3 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -660,8 +674,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string + partition_columns ds/hr + partition_columns.types string:string serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -669,7 +683,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -679,118 +693,6 @@ STAGE PLANS: Table: default.list_bucketing_dynamic_part_n3 Is Table Level Stats: false - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part_n3 - name: default.list_bucketing_dynamic_part_n3 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part_n3 - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part_n3 - name: default.list_bucketing_dynamic_part_n3 - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table list_bucketing_dynamic_part_n3 partition (ds = '2008-04-08', hr) select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08' PREHOOK: type: QUERY @@ -843,7 +745,7 @@ Partition Parameters: numFiles 1 numRows 16 rawDataSize 136 - totalSize 254 + totalSize 235 #### A masked pattern was here #### # Storage Information @@ -881,10 +783,10 @@ Table: list_bucketing_dynamic_part_n3 #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 4 + numFiles 3 numRows 984 rawDataSize 9488 - totalSize 10622 + totalSize 10495 #### A masked pattern was here #### # Storage Information @@ -938,57 +840,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, CAST('val_484' AS STRING) FROM `default`.`list_bucketing_dynamic_part_n3` WHERE `key` = '484' AND `value` = 'val_484' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_dynamic_part_n3 - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=a1 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -1011,7 +870,7 @@ STAGE PLANS: serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 254 + totalSize 235 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -1035,9 +894,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 name: default.list_bucketing_dynamic_part_n3 -#### A masked pattern was here #### Partition - base file name: value=val_484 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -1052,7 +909,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n3 - numFiles 4 + numFiles 3 numRows 984 partition_columns ds/hr partition_columns.types string:string @@ -1060,7 +917,7 @@ STAGE PLANS: serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10622 + totalSize 10495 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -1084,15 +941,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 name: default.list_bucketing_dynamic_part_n3 - Truncated Path -> Alias: - /list_bucketing_dynamic_part_n3/ds=2008-04-08/hr=a1 [list_bucketing_dynamic_part_n3] - /list_bucketing_dynamic_part_n3/ds=2008-04-08/hr=b1/key=484/value=val_484 [list_bucketing_dynamic_part_n3] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_dynamic_part_n3 + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from list_bucketing_dynamic_part_n3 where key = '484' and value = 'val_484' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_dml_7.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out similarity index 52% rename from ql/src/test/results/clientpositive/list_bucket_dml_7.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out index b877e8197b..b517fa046f 100644 --- ql/src/test/results/clientpositive/list_bucket_dml_7.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out @@ -34,216 +34,230 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_dynamic_part + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part + name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} + serialization.ddl struct srcpart { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -274,7 +288,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -333,10 +347,10 @@ Table: list_bucketing_dynamic_part #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 2 + numFiles 1 numRows 16 rawDataSize 136 - totalSize 310 + totalSize 235 #### A masked pattern was here #### # Storage Information @@ -374,10 +388,10 @@ Table: list_bucketing_dynamic_part #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 4 + numFiles 2 numRows 984 rawDataSize 9488 - totalSize 10576 + totalSize 10416 #### A masked pattern was here #### # Storage Information @@ -415,230 +429,230 @@ FROM `default`.`srcpart` WHERE `ds` = '2008-04-08' STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-04-08/ - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/ + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_dynamic_part + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_dynamic_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_dynamic_part + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments + columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### - name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator + File Output Operator bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -660,8 +674,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string + partition_columns ds/hr + partition_columns.types string:string serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -669,7 +683,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -679,118 +693,6 @@ STAGE PLANS: Table: default.list_bucketing_dynamic_part Is Table Level Stats: false - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part - name: default.list_bucketing_dynamic_part - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - input format: org.apache.hadoop.hive.ql.io.rcfile.merge.RCFileBlockMergeInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.list_bucketing_dynamic_part - partition_columns hr - partition_columns.types string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.list_bucketing_dynamic_part - name: default.list_bucketing_dynamic_part - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table list_bucketing_dynamic_part partition (ds = '2008-04-08', hr) select key, value, if(key % 100 == 0, 'a1', 'b1') from srcpart where ds = '2008-04-08' PREHOOK: type: QUERY @@ -843,7 +745,7 @@ Partition Parameters: numFiles 1 numRows 16 rawDataSize 136 - totalSize 254 + totalSize 235 #### A masked pattern was here #### # Storage Information @@ -881,10 +783,10 @@ Table: list_bucketing_dynamic_part #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} - numFiles 3 + numFiles 2 numRows 984 rawDataSize 9488 - totalSize 10520 + totalSize 10416 #### A masked pattern was here #### # Storage Information @@ -938,57 +840,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `key`, CAST('val_484' AS STRING) FROM `default`.`list_bucketing_dynamic_part` WHERE `key` = '484' AND `value` = 'val_484' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: list_bucketing_dynamic_part - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 546 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=a1 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -1011,7 +870,7 @@ STAGE PLANS: serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 254 + totalSize 235 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -1035,9 +894,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part name: default.list_bucketing_dynamic_part -#### A masked pattern was here #### Partition - base file name: hr=b1 input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat partition values: @@ -1052,7 +909,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part - numFiles 3 + numFiles 2 numRows 984 partition_columns ds/hr partition_columns.types string:string @@ -1060,7 +917,7 @@ STAGE PLANS: serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10520 + totalSize 10416 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe @@ -1084,15 +941,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part name: default.list_bucketing_dynamic_part - Truncated Path -> Alias: - /list_bucketing_dynamic_part/ds=2008-04-08/hr=a1 [list_bucketing_dynamic_part] - /list_bucketing_dynamic_part/ds=2008-04-08/hr=b1 [list_bucketing_dynamic_part] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: list_bucketing_dynamic_part + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from list_bucketing_dynamic_part where key = '484' and value = 'val_484' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out new file mode 100644 index 0000000000..5e6e9cc935 --- /dev/null +++ ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out @@ -0,0 +1,853 @@ +PREHOOK: query: create table list_bucketing_static_part_n0 (key String, value String) + partitioned by (ds String, hr String) + skewed by (key) on ('484','103') + stored as DIRECTORIES + STORED AS RCFILE +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@list_bucketing_static_part_n0 +POSTHOOK: query: create table list_bucketing_static_part_n0 (key String, value String) + partitioned by (ds String, hr String) + skewed by (key) on ('484','103') + stored as DIRECTORIES + STORED AS RCFILE +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@list_bucketing_static_part_n0 +PREHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +OPTIMIZED SQL: SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008-04-08 + hr 11 + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n0 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.list_bucketing_static_part_n0 + Is Table Level Stats: false + +PREHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: show partitions list_bucketing_static_part_n0 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: query: show partitions list_bucketing_static_part_n0 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@list_bucketing_static_part_n0 +ds=2008-04-08/hr=11 +PREHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n0 +# col_name data_type comment +key string +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: list_bucketing_static_part_n0 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 3 + numRows 1000 + rawDataSize 9624 + totalSize 10659 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Stored As SubDirectories: Yes +Skewed Columns: [key] +Skewed Values: [[103], [484]] +#### A masked pattern was here #### +Skewed Value to Truncated Path: {[103]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=103, [484]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=484} +Storage Desc Params: + serialization.format 1 +PREHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: query: explain extended +insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +OPTIMIZED SQL: SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `ds` = '2008-04-08' +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-04-08/hr=11/ + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2008-04-08 + hr 11 + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n0 + + Stage: Stage-3 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.list_bucketing_static_part_n0 + Is Table Level Stats: false + +PREHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: query: insert overwrite table list_bucketing_static_part_n0 partition (ds = '2008-04-08', hr = '11') +select key, value from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Output: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).key SIMPLE [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: list_bucketing_static_part_n0 PARTITION(ds=2008-04-08,hr=11).value SIMPLE [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: show partitions list_bucketing_static_part_n0 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: query: show partitions list_bucketing_static_part_n0 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@list_bucketing_static_part_n0 +ds=2008-04-08/hr=11 +PREHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: query: desc formatted list_bucketing_static_part_n0 partition (ds='2008-04-08', hr='11') +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n0 +# col_name data_type comment +key string +value string + +# Partition Information +# col_name data_type comment +ds string +hr string + +# Detailed Partition Information +Partition Value: [2008-04-08, 11] +Database: default +Table: list_bucketing_static_part_n0 +#### A masked pattern was here #### +Partition Parameters: + COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\",\"COLUMN_STATS\":{\"key\":\"true\",\"value\":\"true\"}} + numFiles 3 + numRows 1000 + rawDataSize 9624 + totalSize 10659 +#### A masked pattern was here #### + +# Storage Information +SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat +OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat +Compressed: No +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] +Stored As SubDirectories: Yes +Skewed Columns: [key] +Skewed Values: [[103], [484]] +#### A masked pattern was here #### +Skewed Value to Truncated Path: {[103]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=103, [484]=/list_bucketing_static_part_n0/ds=2008-04-08/hr=11/key=484} +Storage Desc Params: + serialization.format 1 +PREHOOK: query: select count(1) from srcpart where ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +#### A masked pattern was here #### +POSTHOOK: query: select count(1) from srcpart where ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +#### A masked pattern was here #### +1000 +PREHOOK: query: select count(*) from list_bucketing_static_part_n0 +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n0 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from list_bucketing_static_part_n0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n0 +#### A masked pattern was here #### +1000 +PREHOOK: query: explain extended +select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n0 +PREHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: explain extended +select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) AS `$f1`, CAST('2008-04-08' AS STRING) AS `$f2`, CAST('11' AS STRING) AS `$f3` +FROM `default`.`list_bucketing_static_part_n0` +WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '2008-04-08' AND `hr` = '11' +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: + Partition + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + numFiles 3 + numRows 1000 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 9624 + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + totalSize 10659 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.list_bucketing_static_part_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.list_bucketing_static_part_n0 + name: default.list_bucketing_static_part_n0 + Processor Tree: + TableScan + alias: list_bucketing_static_part_n0 + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink + +PREHOOK: query: select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@list_bucketing_static_part_n0 +PREHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +POSTHOOK: query: select * from list_bucketing_static_part_n0 where ds = '2008-04-08' and hr = '11' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: Input: default@list_bucketing_static_part_n0@ds=2008-04-08/hr=11 +#### A masked pattern was here #### +484 val_484 2008-04-08 11 +484 val_484 2008-04-08 11 +PREHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' +PREHOOK: type: QUERY +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: select * from srcpart where ds = '2008-04-08' and key = '484' and value = 'val_484' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +484 val_484 2008-04-08 11 +484 val_484 2008-04-08 12 +PREHOOK: query: drop table list_bucketing_static_part_n0 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@list_bucketing_static_part_n0 +PREHOOK: Output: default@list_bucketing_static_part_n0 +POSTHOOK: query: drop table list_bucketing_static_part_n0 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@list_bucketing_static_part_n0 +POSTHOOK: Output: default@list_bucketing_static_part_n0 diff --git ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out similarity index 63% rename from ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out index 64e4d6a98d..441b71d52c 100644 --- ql/src/test/results/clientpositive/list_bucket_query_multiskew_1.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out @@ -90,57 +90,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0` FROM `default`.`fact_daily` WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 87 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 87 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -187,14 +144,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily - Truncated Path -> Alias: - /fact_daily/ds=1/hr=4/key=484/value=val_484 [fact_daily] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT key FROM fact_daily WHERE ( ds='1' and hr='4') and (key='484' and value= 'val_484') PREHOOK: type: QUERY @@ -221,57 +182,14 @@ OPTIMIZED SQL: SELECT CAST('238' AS STRING) AS `$f0`, CAST('val_238' AS STRING) FROM `default`.`fact_daily` WHERE `key` = '238' AND `value` = 'val_238' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily - filterExpr: ((key = '238') and (value = 'val_238')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '238') and (value = 'val_238')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '238' (type: string), 'val_238' (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_238 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -318,14 +236,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily - Truncated Path -> Alias: - /fact_daily/ds=1/hr=4/key=238/value=val_238 [fact_daily] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily + filterExpr: ((key = '238') and (value = 'val_238')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '238') and (value = 'val_238')) (type: boolean) + Select Operator + expressions: '238' (type: string), 'val_238' (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: SELECT key,value FROM fact_daily WHERE ( ds='1' and hr='4') and (key='238' and value= 'val_238') PREHOOK: type: QUERY @@ -353,57 +275,14 @@ OPTIMIZED SQL: SELECT `key` FROM `default`.`fact_daily` WHERE `value` = '3' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily - filterExpr: ((value = '3') and (ds = '1') and (hr = '4')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (value = '3') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -450,14 +329,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily - Truncated Path -> Alias: - /fact_daily/ds=1/hr=4/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily + filterExpr: ((value = '3') and (ds = '1') and (hr = '4')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (value = '3') (type: boolean) + Select Operator + expressions: key (type: string) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT key FROM fact_daily WHERE ( ds='1' and hr='4') and (value = "3") PREHOOK: type: QUERY @@ -483,57 +366,14 @@ OPTIMIZED SQL: SELECT CAST('495' AS STRING) AS `key`, `value` FROM `default`.`fact_daily` WHERE `key` = '495' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily - filterExpr: ((key = '495') and (ds = '1') and (hr = '4')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '495') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '495' (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -580,14 +420,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily - Truncated Path -> Alias: - /fact_daily/ds=1/hr=4/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily + filterExpr: ((key = '495') and (ds = '1') and (hr = '4')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '495') (type: boolean) + Select Operator + expressions: '495' (type: string), value (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: SELECT key,value FROM fact_daily WHERE ( ds='1' and hr='4') and key = '369' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out similarity index 50% rename from ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out index 98bd32ecbd..4ef05435b7 100644 --- ql/src/test/results/clientpositive/list_bucket_query_multiskew_2.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out @@ -90,106 +90,14 @@ OPTIMIZED SQL: SELECT `key`, CAST('val_484' AS STRING) AS `value` FROM `default`.`fact_daily_n2` WHERE `value` = 'val_484' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n2 - filterExpr: ((value = 'val_484') and (ds = '1') and (hr = '4')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (value = 'val_484') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), 'val_484' (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - hr 4 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - numFiles 3 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n2 - name: default.fact_daily_n2 -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -236,15 +144,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 - Truncated Path -> Alias: - /fact_daily_n2/ds=1/hr=4/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n2] - /fact_daily_n2/ds=1/hr=4/key=484/value=val_484 [fact_daily_n2] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n2 + filterExpr: ((value = 'val_484') and (ds = '1') and (hr = '4')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (value = 'val_484') (type: boolean) + Select Operator + expressions: key (type: string), 'val_484' (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: SELECT key, value FROM fact_daily_n2 WHERE ds='1' and hr='4' and value= 'val_484' PREHOOK: type: QUERY @@ -271,57 +182,14 @@ OPTIMIZED SQL: SELECT CAST('406' AS STRING) AS `$f0` FROM `default`.`fact_daily_n2` WHERE `key` = '406' AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n2 - filterExpr: ((key = '406') and (ds = '1') and (hr = '4')) (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '406') (type: boolean) - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '406' (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -368,14 +236,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 - Truncated Path -> Alias: - /fact_daily_n2/ds=1/hr=4/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n2] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n2 + filterExpr: ((key = '406') and (ds = '1') and (hr = '4')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '406') (type: boolean) + Select Operator + expressions: '406' (type: string) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT key, value FROM fact_daily_n2 WHERE ds='1' and hr='4' and key= '406' PREHOOK: type: QUERY @@ -405,155 +277,14 @@ OPTIMIZED SQL: SELECT `key`, `value` FROM `default`.`fact_daily_n2` WHERE ROW(`key`, `value`) IN (ROW('484', 'val_484'), ROW('238', 'val_238')) AND `ds` = '1' AND `hr` = '4' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n2 - filterExpr: ((struct(key,value)) IN (const struct('484','val_484'), const struct('238','val_238')) and (ds = '1') and (hr = '4')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (struct(key,value)) IN (const struct('484','val_484'), const struct('238','val_238')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - hr 4 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - numFiles 3 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n2 - name: default.fact_daily_n2 -#### A masked pattern was here #### - Partition - base file name: value=val_238 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - hr 4 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - numFiles 3 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.fact_daily_n2 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n2 - name: default.fact_daily_n2 -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -600,16 +331,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 - Truncated Path -> Alias: - /fact_daily_n2/ds=1/hr=4/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n2] - /fact_daily_n2/ds=1/hr=4/key=238/value=val_238 [fact_daily_n2] - /fact_daily_n2/ds=1/hr=4/key=484/value=val_484 [fact_daily_n2] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n2 + filterExpr: ((struct(key,value)) IN (const struct('484','val_484'), const struct('238','val_238')) and (ds = '1') and (hr = '4')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (struct(key,value)) IN (const struct('484','val_484'), const struct('238','val_238')) (type: boolean) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: SELECT key, value FROM fact_daily_n2 WHERE ds='1' and hr='4' and ( (key='484' and value ='val_484') or (key='238' and value= 'val_238')) PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out similarity index 73% rename from ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out index da152c400f..8f3dd7dc25 100644 --- ql/src/test/results/clientpositive/list_bucket_query_multiskew_3.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out @@ -200,57 +200,14 @@ OPTIMIZED SQL: SELECT CAST('145' AS STRING) AS `key`, `value`, CAST('1' AS STRIN FROM `default`.`fact_daily_n3` WHERE `key` = '145' AND `ds` = '1' AND `hr` = '1' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n3 - filterExpr: ((key = '145') and (ds = '1') and (hr = '1')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (key = '145') (type: boolean) - Statistics: Num rows: 2 Data size: 356 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '145' (type: string), value (type: string), '1' (type: string), '1' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 696 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 696 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: hr=1 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -297,14 +254,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 - Truncated Path -> Alias: - /fact_daily_n3/ds=1/hr=1 [fact_daily_n3] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n3 + filterExpr: ((key = '145') and (ds = '1') and (hr = '1')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (key = '145') (type: boolean) + Select Operator + expressions: '145' (type: string), value (type: string), '1' (type: string), '1' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select * from fact_daily_n3 where ds = '1' and hr='1' and key='145' PREHOOK: type: QUERY @@ -365,57 +326,14 @@ OPTIMIZED SQL: SELECT CAST('484' AS STRING) AS `$f0`, CAST('val_484' AS STRING) FROM `default`.`fact_daily_n3` WHERE `key` = '484' AND `value` = 'val_484' AND `ds` = '1' AND `hr` = '2' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n3 - filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '484') and (value = 'val_484')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '484' (type: string), 'val_484' (type: string), '1' (type: string), '2' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -462,14 +380,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 - Truncated Path -> Alias: - /fact_daily_n3/ds=1/hr=2/key=484/value=val_484 [fact_daily_n3] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n3 + filterExpr: ((key = '484') and (value = 'val_484')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '484') and (value = 'val_484')) (type: boolean) + Select Operator + expressions: '484' (type: string), 'val_484' (type: string), '1' (type: string), '2' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: SELECT * FROM fact_daily_n3 WHERE ds='1' and hr='2' and (key='484' and value='val_484') PREHOOK: type: QUERY @@ -498,57 +420,14 @@ OPTIMIZED SQL: SELECT CAST('327' AS STRING) AS `$f0`, CAST('val_327' AS STRING) FROM `default`.`fact_daily_n3` WHERE `key` = '327' AND `value` = 'val_327' AND `ds` = '1' AND `hr` = '3' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n3 - filterExpr: ((key = '327') and (value = 'val_327')) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((key = '327') and (value = 'val_327')) (type: boolean) - Statistics: Num rows: 1 Data size: 178 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '327' (type: string), 'val_327' (type: string), '1' (type: string), '3' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 348 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: value=val_327 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -595,14 +474,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 - Truncated Path -> Alias: - /fact_daily_n3/ds=1/hr=3/key=327/value=val_327 [fact_daily_n3] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n3 + filterExpr: ((key = '327') and (value = 'val_327')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((key = '327') and (value = 'val_327')) (type: boolean) + Select Operator + expressions: '327' (type: string), 'val_327' (type: string), '1' (type: string), '3' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: SELECT * FROM fact_daily_n3 WHERE ds='1' and hr='3' and (key='327' and value='val_327') PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out similarity index 67% rename from ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out index 08f6c82dc9..1bc4d62e5d 100644 --- ql/src/test/results/clientpositive/list_bucket_query_oneskew_1.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out @@ -113,6 +113,7 @@ Sort Columns: [] Skewed Columns: [x] Skewed Values: [[484]] #### A masked pattern was here #### +Skewed Value to Truncated Path: {[484]=/fact_tz/ds=1/x=484, [HIVE_DEFAULT_LIST_BUCKETING_KEY]=/fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME} Storage Desc Params: serialization.format 1 PREHOOK: query: SELECT * FROM fact_daily_n4 WHERE ds='1' @@ -141,57 +142,14 @@ OPTIMIZED SQL: SELECT CAST(484 AS INTEGER) AS `$f0` FROM `default`.`fact_daily_n4` WHERE `x` = 484 AND `ds` = '1' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n4 - filterExpr: ((x = 484) and (ds = '1')) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 484) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 484 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: x=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -235,14 +193,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 - Truncated Path -> Alias: - /fact_tz/ds=1/x=484 [fact_daily_n4] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n4 + filterExpr: ((x = 484) and (ds = '1')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 484) (type: boolean) + Select Operator + expressions: 484 (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT x FROM fact_daily_n4 WHERE ds='1' and x=484 PREHOOK: type: QUERY @@ -269,57 +231,14 @@ OPTIMIZED SQL: SELECT CAST(495 AS INTEGER) AS `$f0` FROM `default`.`fact_daily_n4` WHERE `x` = 495 AND `ds` = '1' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n4 - filterExpr: ((x = 495) and (ds = '1')) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 495) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 495 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -363,14 +282,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 - Truncated Path -> Alias: - /fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n4] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n4 + filterExpr: ((x = 495) and (ds = '1')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 495) (type: boolean) + Select Operator + expressions: 495 (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT x FROM fact_daily_n4 WHERE ds='1' and x=495 PREHOOK: type: QUERY @@ -397,57 +320,14 @@ OPTIMIZED SQL: SELECT CAST(1 AS INTEGER) AS `$f0` FROM `default`.`fact_daily_n4` WHERE `x` = 1 AND `ds` = '1' STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n4 - filterExpr: ((x = 1) and (ds = '1')) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 1) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 1 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -491,14 +371,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 - Truncated Path -> Alias: - /fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n4] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n4 + filterExpr: ((x = 1) and (ds = '1')) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 1) (type: boolean) + Select Operator + expressions: 1 (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT x FROM fact_daily_n4 WHERE ds='1' and x=1 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out similarity index 59% rename from ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out index 972152db05..a60bfe65f8 100644 --- ql/src/test/results/clientpositive/list_bucket_query_oneskew_2.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out @@ -116,6 +116,7 @@ Sort Columns: [] Skewed Columns: [x] Skewed Values: [[484]] #### A masked pattern was here #### +Skewed Value to Truncated Path: {[484]=/fact_tz/ds=1/x=484, [HIVE_DEFAULT_LIST_BUCKETING_KEY]=/fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME} Storage Desc Params: serialization.format 1 PREHOOK: query: SELECT * FROM fact_daily_n5 WHERE ds='1' @@ -144,57 +145,14 @@ OPTIMIZED SQL: SELECT CAST(484 AS INTEGER) AS `$f0` FROM `default`.`fact_daily_n5` WHERE `ds` = '1' AND `x` = 484 STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n5 - filterExpr: ((ds = '1') and (x = 484)) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 484) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 484 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: x=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -238,14 +196,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 - Truncated Path -> Alias: - /fact_tz/ds=1/x=484 [fact_daily_n5] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n5 + filterExpr: ((ds = '1') and (x = 484)) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 484) (type: boolean) + Select Operator + expressions: 484 (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: select x from (select * from fact_daily_n5 where ds = '1') subq where x = 484 PREHOOK: type: QUERY @@ -272,57 +234,14 @@ OPTIMIZED SQL: SELECT CAST(484 AS INTEGER) AS `x1`, `y` AS `y1` FROM `default`.`fact_daily_n5` WHERE `ds` = '1' AND `x` = 484 STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n5 - filterExpr: ((ds = '1') and (x = 484)) (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 484) (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 484 (type: int), y (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types int:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: x=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -366,14 +285,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 - Truncated Path -> Alias: - /fact_tz/ds=1/x=484 [fact_daily_n5] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n5 + filterExpr: ((ds = '1') and (x = 484)) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 484) (type: boolean) + Select Operator + expressions: 484 (type: int), y (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select x1, y1 from(select x as x1, y as y1 from fact_daily_n5 where ds ='1') subq where x1 = 484 PREHOOK: type: QUERY @@ -406,123 +329,132 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n5 - filterExpr: ((ds = '1') and (x = 484)) (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 484) (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: y (type: string) - outputColumnNames: y - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: count() - keys: y (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: + Tez #### A masked pattern was here #### - Path -> Partition: + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### - Partition - base file name: x=484 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - properties: - bucket_count -1 - column.name.delimiter , - columns x,y - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.fact_daily_n5 - numFiles 2 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact_daily_n5 + filterExpr: ((ds = '1') and (x = 484)) (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 484) (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: y (type: string) + outputColumnNames: y + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: count() + keys: y (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: x=484 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 1 + properties: + bucket_count -1 + column.name.delimiter , + columns x,y + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.fact_daily_n5 + numFiles 2 + partition_columns ds + partition_columns.types string + serialization.ddl struct fact_daily_n5 { i32 x, string y} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 24 #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - EXTERNAL TRUE - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns x,y - columns.comments - columns.types int:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + EXTERNAL TRUE + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns x,y + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.fact_daily_n5 + partition_columns ds + partition_columns.types string + serialization.ddl struct fact_daily_n5 { i32 x, string y} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.fact_daily_n5 + name: default.fact_daily_n5 + Truncated Path -> Alias: + /fact_tz/ds=1/x=484 [fact_daily_n5] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - name default.fact_daily_n5 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n5 - name: default.fact_daily_n5 - Truncated Path -> Alias: - /fact_tz/ds=1/x=484 [fact_daily_n5] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 188 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types string:bigint - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types string:bigint + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator @@ -561,125 +493,134 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n5 - filterExpr: ((ds = '1') and (x = 484)) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x = 484) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: count() - keys: true (type: boolean) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: boolean) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - tag: -1 - value expressions: _col1 (type: bigint) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: + Tez #### A masked pattern was here #### - Path -> Partition: + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### - Partition - base file name: x=484 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - properties: - bucket_count -1 - column.name.delimiter , - columns x,y - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.fact_daily_n5 - numFiles 2 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - EXTERNAL TRUE - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns x,y - columns.comments - columns.types int:string -#### A masked pattern was here #### - name default.fact_daily_n5 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n5 - name: default.fact_daily_n5 - Truncated Path -> Alias: - /fact_tz/ds=1/x=484 [$hdt$_0:fact_daily_n5] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: boolean) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: 484 (type: int), _col1 (type: bigint) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: fact_daily_n5 + filterExpr: ((ds = '1') and (x = 484)) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x = 484) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + Select Operator + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: count() + keys: true (type: boolean) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: boolean) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + tag: -1 + value expressions: _col1 (type: bigint) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: x=484 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 1 properties: - bucketing_version -1 - columns _col0,_col1 - columns.types int:bigint - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true + bucket_count -1 + column.name.delimiter , + columns x,y + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.fact_daily_n5 + numFiles 2 + partition_columns ds + partition_columns.types string + serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 24 +#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + EXTERNAL TRUE + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns x,y + columns.comments + columns.types int:string +#### A masked pattern was here #### + name default.fact_daily_n5 + partition_columns ds + partition_columns.types string + serialization.ddl struct fact_daily_n5 { i32 x, string y} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.fact_daily_n5 + name: default.fact_daily_n5 + Truncated Path -> Alias: + /fact_tz/ds=1/x=484 [fact_daily_n5] + Reducer 2 + Execution mode: vectorized, llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: count(VALUE._col0) + keys: KEY._col0 (type: boolean) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: 484 (type: int), _col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 4 Basic stats: PARTIAL Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types int:bigint + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out similarity index 61% rename from ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out rename to ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out index 234d26f6b0..1bfbfd3f65 100644 --- ql/src/test/results/clientpositive/list_bucket_query_oneskew_3.q.out +++ ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out @@ -132,6 +132,7 @@ Sort Columns: [] Skewed Columns: [x] Skewed Values: [[238], [484]] #### A masked pattern was here #### +Skewed Value to Truncated Path: {[238]=/fact_tz/ds=1/x=238, [484]=/fact_tz/ds=1/x=484, [HIVE_DEFAULT_LIST_BUCKETING_KEY]=/fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME} Storage Desc Params: serialization.format 1 PREHOOK: query: SELECT * FROM fact_daily_n0 WHERE ds='1' @@ -164,149 +165,14 @@ OPTIMIZED SQL: SELECT `x` FROM `default`.`fact_daily_n0` WHERE `ds` = '1' AND `x` <> 86 STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: fact_daily_n0 - filterExpr: ((ds = '1') and (x <> 86)) (type: boolean) - Statistics: Num rows: 2 Data size: 8 Basic stats: PARTIAL Column stats: NONE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (x <> 86) (type: boolean) - Statistics: Num rows: 2 Data size: 8 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: x (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 8 Basic stats: PARTIAL Column stats: NONE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 8 Basic stats: PARTIAL Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0 - columns.types int - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - properties: - bucket_count -1 - column.name.delimiter , - columns x,y,z - columns.comments - columns.types int:string:string -#### A masked pattern was here #### - name default.fact_daily_n0 - numFiles 3 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 117 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - EXTERNAL TRUE - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns x,y,z - columns.comments - columns.types int:string:string -#### A masked pattern was here #### - name default.fact_daily_n0 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n0 - name: default.fact_daily_n0 -#### A masked pattern was here #### - Partition - base file name: x=238 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 1 - properties: - bucket_count -1 - column.name.delimiter , - columns x,y,z - columns.comments - columns.types int:string:string -#### A masked pattern was here #### - name default.fact_daily_n0 - numFiles 3 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 117 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - EXTERNAL TRUE - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns x,y,z - columns.comments - columns.types int:string:string -#### A masked pattern was here #### - name default.fact_daily_n0 - partition_columns ds - partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.fact_daily_n0 - name: default.fact_daily_n0 -#### A masked pattern was here #### + Stage: Stage-0 + Fetch Operator + limit: -1 + Partition Description: Partition - base file name: x=484 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat partition values: @@ -350,16 +216,18 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n0 name: default.fact_daily_n0 - Truncated Path -> Alias: - /fact_tz/ds=1/HIVE_DEFAULT_LIST_BUCKETING_DIR_NAME [fact_daily_n0] - /fact_tz/ds=1/x=238 [fact_daily_n0] - /fact_tz/ds=1/x=484 [fact_daily_n0] - - Stage: Stage-0 - Fetch Operator - limit: -1 Processor Tree: - ListSink + TableScan + alias: fact_daily_n0 + filterExpr: ((ds = '1') and (x <> 86)) (type: boolean) + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (x <> 86) (type: boolean) + Select Operator + expressions: x (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: SELECT x FROM fact_daily_n0 WHERE ds='1' and not (x = 86) PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/literal_decimal.q.out ql/src/test/results/clientpositive/llap/literal_decimal.q.out similarity index 84% rename from ql/src/test/results/clientpositive/literal_decimal.q.out rename to ql/src/test/results/clientpositive/llap/literal_decimal.q.out index 64112ecd91..cefc9e3696 100644 --- ql/src/test/results/clientpositive/literal_decimal.q.out +++ ql/src/test/results/clientpositive/llap/literal_decimal.q.out @@ -16,14 +16,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: -1 (type: decimal(1,0)), 0 (type: decimal(1,0)), 1 (type: decimal(1,0)), 3.14 (type: decimal(3,2)), -3.14 (type: decimal(3,2)), 99999999999999999 (type: decimal(17,0)), 99999999999999999.9999999999999 (type: decimal(30,13)), null (type: decimal(1,0)) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 500 Data size: 392112 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 896 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT -1BD, 0BD, 1BD, 3.14BD, -3.14BD, 99999999999999999BD, 99999999999999999.9999999999999BD, 1E99BD FROM src LIMIT 1 diff --git ql/src/test/results/clientpositive/literal_double.q.out ql/src/test/results/clientpositive/llap/literal_double.q.out similarity index 83% rename from ql/src/test/results/clientpositive/literal_double.q.out rename to ql/src/test/results/clientpositive/llap/literal_double.q.out index 81459c9926..599ecfef07 100644 --- ql/src/test/results/clientpositive/literal_double.q.out +++ ql/src/test/results/clientpositive/llap/literal_double.q.out @@ -16,14 +16,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 3.14 (type: decimal(3,2)), -3.14 (type: decimal(3,2)), 3.14E8D (type: double), 3.14E-8D (type: double), -3.14E8D (type: double), -3.14E-8D (type: double), 3.14E8D (type: double), 3.14E8D (type: double), 3.14E-8D (type: double) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 500 Data size: 140000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 280 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT 3.14, -3.14, 3.14e8, 3.14e-8, -3.14e8, -3.14e-8, 3.14e+8, 3.14E8, 3.14E-8 FROM src LIMIT 1 diff --git ql/src/test/results/clientpositive/literal_ints.q.out ql/src/test/results/clientpositive/llap/literal_ints.q.out similarity index 78% rename from ql/src/test/results/clientpositive/literal_ints.q.out rename to ql/src/test/results/clientpositive/llap/literal_ints.q.out index 9867e09b83..66b197f4a9 100644 --- ql/src/test/results/clientpositive/literal_ints.q.out +++ ql/src/test/results/clientpositive/llap/literal_ints.q.out @@ -16,14 +16,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 100 (type: int), 100Y (type: tinyint), 100S (type: smallint), 100L (type: bigint) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 10000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT 100, 100Y, 100S, 100L FROM src LIMIT 1 diff --git ql/src/test/results/clientpositive/literal_string.q.out ql/src/test/results/clientpositive/llap/literal_string.q.out similarity index 88% rename from ql/src/test/results/clientpositive/literal_string.q.out rename to ql/src/test/results/clientpositive/llap/literal_string.q.out index c5bfd75fc1..5fa5120e35 100644 --- ql/src/test/results/clientpositive/literal_string.q.out +++ ql/src/test/results/clientpositive/llap/literal_string.q.out @@ -26,14 +26,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string), 'facebook' (type: string) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 - Statistics: Num rows: 500 Data size: 460000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: SELECT 'face''book', 'face' 'book', 'face' diff --git ql/src/test/results/clientpositive/llap/llap_reader.q.out ql/src/test/results/clientpositive/llap/llap_reader.q.out new file mode 100644 index 0000000000..c2fdd18482 --- /dev/null +++ ql/src/test/results/clientpositive/llap/llap_reader.q.out @@ -0,0 +1,225 @@ +PREHOOK: query: CREATE TABLE test_n7(f1 int, f2 int, f3 int) stored as orc +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_n7 +PREHOOK: query: INSERT INTO TABLE test_n7 VALUES (1,1,1), (2,2,2), (3,3,3), (4,4,4) +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@test_n7 +PREHOOK: query: ALTER TABLE test_n7 CHANGE f1 f1 bigint +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_n7 +PREHOOK: Output: default@test_n7 +PREHOOK: query: ALTER TABLE test_n7 CHANGE f2 f2 bigint +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_n7 +PREHOOK: Output: default@test_n7 +PREHOOK: query: ALTER TABLE test_n7 CHANGE f3 f3 bigint +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_n7 +PREHOOK: Output: default@test_n7 +PREHOOK: query: SELECT count(f1) FROM test_n7 GROUP BY f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_n7 +#### A masked pattern was here #### +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 2 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 4 + RECORDS_OUT_0: 4 + RECORDS_OUT_INTERMEDIATE_Map_1: 4 + RECORDS_OUT_INTERMEDIATE_Reducer_2: 0 + RECORDS_OUT_OPERATOR_FS_10: 4 + RECORDS_OUT_OPERATOR_GBY_8: 4 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_RS_7: 4 + RECORDS_OUT_OPERATOR_SEL_6: 4 + RECORDS_OUT_OPERATOR_SEL_9: 4 + RECORDS_OUT_OPERATOR_TS_0: 4 + TOTAL_TABLE_ROWS_WRITTEN: 0 +Stage-1 LLAP IO COUNTERS: + ALLOCATED_BYTES: 786432 + ALLOCATED_USED_BYTES: 33 + CACHE_HIT_BYTES: 0 + CACHE_MISS_BYTES: 42 + METADATA_CACHE_MISS: 2 + NUM_DECODED_BATCHES: 1 + NUM_VECTOR_BATCHES: 1 + ROWS_EMITTED: 4 + SELECTED_ROWGROUPS: 1 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +1 +1 +1 +1 +PREHOOK: query: SELECT count(f1) FROM test_n7 GROUP BY f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_n7 +#### A masked pattern was here #### +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 2 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 4 + RECORDS_OUT_0: 4 + RECORDS_OUT_INTERMEDIATE_Map_1: 4 + RECORDS_OUT_INTERMEDIATE_Reducer_2: 0 + RECORDS_OUT_OPERATOR_FS_10: 4 + RECORDS_OUT_OPERATOR_GBY_8: 4 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_RS_7: 4 + RECORDS_OUT_OPERATOR_SEL_6: 4 + RECORDS_OUT_OPERATOR_SEL_9: 4 + RECORDS_OUT_OPERATOR_TS_0: 4 + TOTAL_TABLE_ROWS_WRITTEN: 0 +Stage-1 LLAP IO COUNTERS: + CACHE_HIT_BYTES: 42 + CACHE_MISS_BYTES: 0 + METADATA_CACHE_HIT: 2 + NUM_DECODED_BATCHES: 1 + NUM_VECTOR_BATCHES: 1 + ROWS_EMITTED: 4 + SELECTED_ROWGROUPS: 1 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +1 +1 +1 +1 +PREHOOK: query: CREATE TABLE test_bigint(f1 bigint, f2 bigint, f3 bigint) stored as orc +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_bigint +PREHOOK: query: INSERT OVERWRITE TABLE test_bigint select * from test_n7 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_n7 +PREHOOK: Output: default@test_bigint +PREHOOK: query: ALTER TABLE test_bigint CHANGE f1 f1 double +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_bigint +PREHOOK: Output: default@test_bigint +PREHOOK: query: ALTER TABLE test_bigint CHANGE f2 f2 double +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_bigint +PREHOOK: Output: default@test_bigint +PREHOOK: query: ALTER TABLE test_bigint CHANGE f3 f3 double +PREHOOK: type: ALTERTABLE_RENAMECOL +PREHOOK: Input: default@test_bigint +PREHOOK: Output: default@test_bigint +PREHOOK: query: select count(f1) from test_bigint group by f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_bigint +#### A masked pattern was here #### +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 2 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 4 + RECORDS_OUT_0: 4 + RECORDS_OUT_INTERMEDIATE_Map_1: 4 + RECORDS_OUT_INTERMEDIATE_Reducer_2: 0 + RECORDS_OUT_OPERATOR_FS_10: 4 + RECORDS_OUT_OPERATOR_GBY_8: 4 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_RS_7: 4 + RECORDS_OUT_OPERATOR_SEL_6: 4 + RECORDS_OUT_OPERATOR_SEL_9: 4 + RECORDS_OUT_OPERATOR_TS_0: 4 + TOTAL_TABLE_ROWS_WRITTEN: 0 +Stage-1 LLAP IO COUNTERS: + METADATA_CACHE_MISS: 1 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +1 +1 +1 +1 +PREHOOK: query: select count(f1) from test_bigint group by f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_bigint +#### A masked pattern was here #### +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 2 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 4 + RECORDS_OUT_0: 4 + RECORDS_OUT_INTERMEDIATE_Map_1: 4 + RECORDS_OUT_INTERMEDIATE_Reducer_2: 0 + RECORDS_OUT_OPERATOR_FS_10: 4 + RECORDS_OUT_OPERATOR_GBY_8: 4 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_RS_7: 4 + RECORDS_OUT_OPERATOR_SEL_6: 4 + RECORDS_OUT_OPERATOR_SEL_9: 4 + RECORDS_OUT_OPERATOR_TS_0: 4 + TOTAL_TABLE_ROWS_WRITTEN: 0 +Stage-1 LLAP IO COUNTERS: + METADATA_CACHE_HIT: 1 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 1 + RAW_INPUT_SPLITS_Map_1: 1 +1 +1 +1 +1 +PREHOOK: query: CREATE TABLE test_acid_n0 (f1 int, f2 int, val string) clustered by (val) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true') +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_acid_n0 +PREHOOK: query: INSERT INTO TABLE test_acid_n0 VALUES (1,1,'b1'), (2,2,'b2'), (3,3,'b3'), (4,4,'b4') +PREHOOK: type: QUERY +PREHOOK: Input: _dummy_database@_dummy_table +PREHOOK: Output: default@test_acid_n0 +PREHOOK: query: SELECT count(f1) FROM test_acid_n0 GROUP BY f1 +PREHOOK: type: QUERY +PREHOOK: Input: default@test_acid_n0 +#### A masked pattern was here #### +Stage-1 FILE SYSTEM COUNTERS: +Stage-1 HIVE COUNTERS: + CREATED_FILES: 2 + DESERIALIZE_ERRORS: 0 + RECORDS_IN_Map_1: 4 + RECORDS_OUT_0: 4 + RECORDS_OUT_INTERMEDIATE_Map_1: 4 + RECORDS_OUT_INTERMEDIATE_Reducer_2: 0 + RECORDS_OUT_OPERATOR_FS_10: 4 + RECORDS_OUT_OPERATOR_GBY_8: 4 + RECORDS_OUT_OPERATOR_MAP_0: 0 + RECORDS_OUT_OPERATOR_RS_7: 4 + RECORDS_OUT_OPERATOR_SEL_6: 4 + RECORDS_OUT_OPERATOR_SEL_9: 4 + RECORDS_OUT_OPERATOR_TS_0: 4 + TOTAL_TABLE_ROWS_WRITTEN: 0 +Stage-1 LLAP IO COUNTERS: + ALLOCATED_BYTES: 196608 + ALLOCATED_USED_BYTES: 65 + CACHE_HIT_BYTES: 0 + CACHE_MISS_BYTES: 83 + METADATA_CACHE_MISS: 4 + NUM_DECODED_BATCHES: 2 + NUM_VECTOR_BATCHES: 2 + ROWS_EMITTED: 4 + SELECTED_ROWGROUPS: 2 +Stage-1 INPUT COUNTERS: + GROUPED_INPUT_SPLITS_Map_1: 1 + INPUT_DIRECTORIES_Map_1: 1 + INPUT_FILES_Map_1: 2 + RAW_INPUT_SPLITS_Map_1: 2 +1 +1 +1 +1 diff --git ql/src/test/results/clientpositive/load_dyn_part10.q.out ql/src/test/results/clientpositive/llap/load_dyn_part10.q.out similarity index 87% rename from ql/src/test/results/clientpositive/load_dyn_part10.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part10.q.out index e907b0cffb..f9ced8e400 100644 --- ql/src/test/results/clientpositive/load_dyn_part10.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part10.q.out @@ -50,65 +50,79 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds > '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part10 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-12-31' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds > '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part10 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-12-31' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -123,7 +137,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part10 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: diff --git ql/src/test/results/clientpositive/load_dyn_part11.q.out ql/src/test/results/clientpositive/llap/load_dyn_part11.q.out similarity index 100% rename from ql/src/test/results/clientpositive/load_dyn_part11.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part11.q.out index 92d7162961..b7031b6cf2 100644 --- ql/src/test/results/clientpositive/load_dyn_part11.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part11.q.out @@ -64,1006 +64,1006 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@nzhang_part_n0 POSTHOOK: Input: default@nzhang_part_n0@ds=2010-03-03/hr=11 #### A masked pattern was here #### -238 val_238 2010-03-03 11 -97 val_97 2010-03-03 11 -200 val_200 2010-03-03 11 -400 val_400 2010-03-03 11 -403 val_403 2010-03-03 11 -169 val_169 2010-03-03 11 -90 val_90 2010-03-03 11 -126 val_126 2010-03-03 11 -222 val_222 2010-03-03 11 -477 val_477 2010-03-03 11 -414 val_414 2010-03-03 11 -194 val_194 2010-03-03 11 -307 val_307 2010-03-03 11 -348 val_348 2010-03-03 11 -152 val_152 2010-03-03 11 -448 val_448 2010-03-03 11 -37 val_37 2010-03-03 11 -28 val_28 2010-03-03 11 -84 val_84 2010-03-03 11 -315 val_315 2010-03-03 11 -469 val_469 2010-03-03 11 -97 val_97 2010-03-03 11 -344 val_344 2010-03-03 11 -281 val_281 2010-03-03 11 -183 val_183 2010-03-03 11 -273 val_273 2010-03-03 11 -18 val_18 2010-03-03 11 -167 val_167 2010-03-03 11 -348 val_348 2010-03-03 11 -285 val_285 2010-03-03 11 -186 val_186 2010-03-03 11 -362 val_362 2010-03-03 11 -458 val_458 2010-03-03 11 -146 val_146 2010-03-03 11 -498 val_498 2010-03-03 11 -341 val_341 2010-03-03 11 -9 val_9 2010-03-03 11 -298 val_298 2010-03-03 11 +0 val_0 2010-03-03 11 +0 val_0 2010-03-03 11 +0 val_0 2010-03-03 11 +0 val_0 2010-03-03 11 +0 val_0 2010-03-03 11 +0 val_0 2010-03-03 11 +10 val_10 2010-03-03 11 +10 val_10 2010-03-03 11 100 val_100 2010-03-03 11 -492 val_492 2010-03-03 11 -462 val_462 2010-03-03 11 -18 val_18 2010-03-03 11 -379 val_379 2010-03-03 11 -384 val_384 2010-03-03 11 -67 val_67 2010-03-03 11 -134 val_134 2010-03-03 11 -26 val_26 2010-03-03 11 -256 val_256 2010-03-03 11 -384 val_384 2010-03-03 11 -407 val_407 2010-03-03 11 -421 val_421 2010-03-03 11 -401 val_401 2010-03-03 11 -375 val_375 2010-03-03 11 -454 val_454 2010-03-03 11 -189 val_189 2010-03-03 11 -175 val_175 2010-03-03 11 +100 val_100 2010-03-03 11 +100 val_100 2010-03-03 11 +100 val_100 2010-03-03 11 +103 val_103 2010-03-03 11 +103 val_103 2010-03-03 11 +103 val_103 2010-03-03 11 +103 val_103 2010-03-03 11 +104 val_104 2010-03-03 11 +104 val_104 2010-03-03 11 +104 val_104 2010-03-03 11 +104 val_104 2010-03-03 11 +105 val_105 2010-03-03 11 +105 val_105 2010-03-03 11 +11 val_11 2010-03-03 11 +11 val_11 2010-03-03 11 +111 val_111 2010-03-03 11 +111 val_111 2010-03-03 11 +113 val_113 2010-03-03 11 +113 val_113 2010-03-03 11 +113 val_113 2010-03-03 11 +113 val_113 2010-03-03 11 +114 val_114 2010-03-03 11 +114 val_114 2010-03-03 11 +116 val_116 2010-03-03 11 +116 val_116 2010-03-03 11 +118 val_118 2010-03-03 11 +118 val_118 2010-03-03 11 +118 val_118 2010-03-03 11 +118 val_118 2010-03-03 11 +119 val_119 2010-03-03 11 +119 val_119 2010-03-03 11 +119 val_119 2010-03-03 11 +119 val_119 2010-03-03 11 +119 val_119 2010-03-03 11 +119 val_119 2010-03-03 11 +12 val_12 2010-03-03 11 +12 val_12 2010-03-03 11 +12 val_12 2010-03-03 11 +12 val_12 2010-03-03 11 +120 val_120 2010-03-03 11 +120 val_120 2010-03-03 11 +120 val_120 2010-03-03 11 +120 val_120 2010-03-03 11 +125 val_125 2010-03-03 11 +125 val_125 2010-03-03 11 +125 val_125 2010-03-03 11 +125 val_125 2010-03-03 11 +126 val_126 2010-03-03 11 +126 val_126 2010-03-03 11 +128 val_128 2010-03-03 11 +128 val_128 2010-03-03 11 +128 val_128 2010-03-03 11 +128 val_128 2010-03-03 11 +128 val_128 2010-03-03 11 +128 val_128 2010-03-03 11 +129 val_129 2010-03-03 11 +129 val_129 2010-03-03 11 +129 val_129 2010-03-03 11 +129 val_129 2010-03-03 11 +131 val_131 2010-03-03 11 +131 val_131 2010-03-03 11 133 val_133 2010-03-03 11 -406 val_406 2010-03-03 11 -233 val_233 2010-03-03 11 -462 val_462 2010-03-03 11 -214 val_214 2010-03-03 11 -172 val_172 2010-03-03 11 -353 val_353 2010-03-03 11 +133 val_133 2010-03-03 11 +134 val_134 2010-03-03 11 +134 val_134 2010-03-03 11 +134 val_134 2010-03-03 11 +134 val_134 2010-03-03 11 136 val_136 2010-03-03 11 -83 val_83 2010-03-03 11 -480 val_480 2010-03-03 11 -265 val_265 2010-03-03 11 -249 val_249 2010-03-03 11 -207 val_207 2010-03-03 11 -460 val_460 2010-03-03 11 -493 val_493 2010-03-03 11 -333 val_333 2010-03-03 11 -317 val_317 2010-03-03 11 -310 val_310 2010-03-03 11 -468 val_468 2010-03-03 11 -178 val_178 2010-03-03 11 -478 val_478 2010-03-03 11 -230 val_230 2010-03-03 11 -277 val_277 2010-03-03 11 -325 val_325 2010-03-03 11 -323 val_323 2010-03-03 11 -443 val_443 2010-03-03 11 -169 val_169 2010-03-03 11 -429 val_429 2010-03-03 11 -120 val_120 2010-03-03 11 -444 val_444 2010-03-03 11 -199 val_199 2010-03-03 11 -417 val_417 2010-03-03 11 -305 val_305 2010-03-03 11 -479 val_479 2010-03-03 11 -248 val_248 2010-03-03 11 -360 val_360 2010-03-03 11 -439 val_439 2010-03-03 11 -237 val_237 2010-03-03 11 -491 val_491 2010-03-03 11 -200 val_200 2010-03-03 11 -414 val_414 2010-03-03 11 -119 val_119 2010-03-03 11 -438 val_438 2010-03-03 11 +136 val_136 2010-03-03 11 +137 val_137 2010-03-03 11 +137 val_137 2010-03-03 11 +137 val_137 2010-03-03 11 +137 val_137 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +138 val_138 2010-03-03 11 +143 val_143 2010-03-03 11 +143 val_143 2010-03-03 11 +145 val_145 2010-03-03 11 +145 val_145 2010-03-03 11 +146 val_146 2010-03-03 11 +146 val_146 2010-03-03 11 +146 val_146 2010-03-03 11 +146 val_146 2010-03-03 11 +149 val_149 2010-03-03 11 +149 val_149 2010-03-03 11 +149 val_149 2010-03-03 11 +149 val_149 2010-03-03 11 +15 val_15 2010-03-03 11 +15 val_15 2010-03-03 11 +15 val_15 2010-03-03 11 +15 val_15 2010-03-03 11 +150 val_150 2010-03-03 11 +150 val_150 2010-03-03 11 +152 val_152 2010-03-03 11 +152 val_152 2010-03-03 11 +152 val_152 2010-03-03 11 +152 val_152 2010-03-03 11 +153 val_153 2010-03-03 11 +153 val_153 2010-03-03 11 +155 val_155 2010-03-03 11 +155 val_155 2010-03-03 11 +156 val_156 2010-03-03 11 +156 val_156 2010-03-03 11 +157 val_157 2010-03-03 11 +157 val_157 2010-03-03 11 +158 val_158 2010-03-03 11 +158 val_158 2010-03-03 11 +160 val_160 2010-03-03 11 +160 val_160 2010-03-03 11 +162 val_162 2010-03-03 11 +162 val_162 2010-03-03 11 +163 val_163 2010-03-03 11 163 val_163 2010-03-03 11 -70 val_70 2010-03-03 11 -104 val_104 2010-03-03 11 -255 val_255 2010-03-03 11 -351 val_351 2010-03-03 11 -24 val_24 2010-03-03 11 -291 val_291 2010-03-03 11 -480 val_480 2010-03-03 11 -397 val_397 2010-03-03 11 -70 val_70 2010-03-03 11 -5 val_5 2010-03-03 11 -382 val_382 2010-03-03 11 -187 val_187 2010-03-03 11 -424 val_424 2010-03-03 11 164 val_164 2010-03-03 11 -431 val_431 2010-03-03 11 -125 val_125 2010-03-03 11 -298 val_298 2010-03-03 11 -478 val_478 2010-03-03 11 -454 val_454 2010-03-03 11 -431 val_431 2010-03-03 11 164 val_164 2010-03-03 11 -217 val_217 2010-03-03 11 -201 val_201 2010-03-03 11 -396 val_396 2010-03-03 11 -12 val_12 2010-03-03 11 -424 val_424 2010-03-03 11 -348 val_348 2010-03-03 11 -262 val_262 2010-03-03 11 -203 val_203 2010-03-03 11 -90 val_90 2010-03-03 11 -258 val_258 2010-03-03 11 -114 val_114 2010-03-03 11 -401 val_401 2010-03-03 11 -406 val_406 2010-03-03 11 -190 val_190 2010-03-03 11 -409 val_409 2010-03-03 11 -406 val_406 2010-03-03 11 -257 val_257 2010-03-03 11 -105 val_105 2010-03-03 11 -53 val_53 2010-03-03 11 -483 val_483 2010-03-03 11 -403 val_403 2010-03-03 11 +164 val_164 2010-03-03 11 +164 val_164 2010-03-03 11 +165 val_165 2010-03-03 11 +165 val_165 2010-03-03 11 +165 val_165 2010-03-03 11 +165 val_165 2010-03-03 11 +166 val_166 2010-03-03 11 +166 val_166 2010-03-03 11 +167 val_167 2010-03-03 11 +167 val_167 2010-03-03 11 +167 val_167 2010-03-03 11 +167 val_167 2010-03-03 11 +167 val_167 2010-03-03 11 +167 val_167 2010-03-03 11 +168 val_168 2010-03-03 11 +168 val_168 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +169 val_169 2010-03-03 11 +17 val_17 2010-03-03 11 +17 val_17 2010-03-03 11 +170 val_170 2010-03-03 11 +170 val_170 2010-03-03 11 +172 val_172 2010-03-03 11 +172 val_172 2010-03-03 11 +172 val_172 2010-03-03 11 +172 val_172 2010-03-03 11 +174 val_174 2010-03-03 11 +174 val_174 2010-03-03 11 +174 val_174 2010-03-03 11 +174 val_174 2010-03-03 11 175 val_175 2010-03-03 11 -366 val_366 2010-03-03 11 -466 val_466 2010-03-03 11 -104 val_104 2010-03-03 11 -335 val_335 2010-03-03 11 -321 val_321 2010-03-03 11 +175 val_175 2010-03-03 11 +175 val_175 2010-03-03 11 +175 val_175 2010-03-03 11 +176 val_176 2010-03-03 11 +176 val_176 2010-03-03 11 +176 val_176 2010-03-03 11 +176 val_176 2010-03-03 11 +177 val_177 2010-03-03 11 +177 val_177 2010-03-03 11 +178 val_178 2010-03-03 11 +178 val_178 2010-03-03 11 +179 val_179 2010-03-03 11 +179 val_179 2010-03-03 11 +179 val_179 2010-03-03 11 +179 val_179 2010-03-03 11 +18 val_18 2010-03-03 11 +18 val_18 2010-03-03 11 +18 val_18 2010-03-03 11 +18 val_18 2010-03-03 11 +180 val_180 2010-03-03 11 +180 val_180 2010-03-03 11 +181 val_181 2010-03-03 11 +181 val_181 2010-03-03 11 +183 val_183 2010-03-03 11 +183 val_183 2010-03-03 11 +186 val_186 2010-03-03 11 +186 val_186 2010-03-03 11 +187 val_187 2010-03-03 11 +187 val_187 2010-03-03 11 +187 val_187 2010-03-03 11 +187 val_187 2010-03-03 11 +187 val_187 2010-03-03 11 +187 val_187 2010-03-03 11 +189 val_189 2010-03-03 11 +189 val_189 2010-03-03 11 +19 val_19 2010-03-03 11 +19 val_19 2010-03-03 11 +190 val_190 2010-03-03 11 +190 val_190 2010-03-03 11 +191 val_191 2010-03-03 11 +191 val_191 2010-03-03 11 +191 val_191 2010-03-03 11 +191 val_191 2010-03-03 11 +192 val_192 2010-03-03 11 +192 val_192 2010-03-03 11 193 val_193 2010-03-03 11 -44 val_44 2010-03-03 11 -80 val_80 2010-03-03 11 -235 val_235 2010-03-03 11 -331 val_331 2010-03-03 11 -283 val_283 2010-03-03 11 -35 val_35 2010-03-03 11 +193 val_193 2010-03-03 11 +193 val_193 2010-03-03 11 +193 val_193 2010-03-03 11 +193 val_193 2010-03-03 11 +193 val_193 2010-03-03 11 +194 val_194 2010-03-03 11 +194 val_194 2010-03-03 11 +195 val_195 2010-03-03 11 +195 val_195 2010-03-03 11 +195 val_195 2010-03-03 11 +195 val_195 2010-03-03 11 +196 val_196 2010-03-03 11 +196 val_196 2010-03-03 11 +197 val_197 2010-03-03 11 +197 val_197 2010-03-03 11 +197 val_197 2010-03-03 11 +197 val_197 2010-03-03 11 +199 val_199 2010-03-03 11 +199 val_199 2010-03-03 11 +199 val_199 2010-03-03 11 +199 val_199 2010-03-03 11 +199 val_199 2010-03-03 11 +199 val_199 2010-03-03 11 2 val_2 2010-03-03 11 -280 val_280 2010-03-03 11 -463 val_463 2010-03-03 11 -469 val_469 2010-03-03 11 -229 val_229 2010-03-03 11 -316 val_316 2010-03-03 11 +2 val_2 2010-03-03 11 +20 val_20 2010-03-03 11 +20 val_20 2010-03-03 11 +200 val_200 2010-03-03 11 +200 val_200 2010-03-03 11 +200 val_200 2010-03-03 11 +200 val_200 2010-03-03 11 +201 val_201 2010-03-03 11 +201 val_201 2010-03-03 11 202 val_202 2010-03-03 11 -432 val_432 2010-03-03 11 -467 val_467 2010-03-03 11 -128 val_128 2010-03-03 11 -438 val_438 2010-03-03 11 -244 val_244 2010-03-03 11 -5 val_5 2010-03-03 11 -191 val_191 2010-03-03 11 -288 val_288 2010-03-03 11 -401 val_401 2010-03-03 11 -480 val_480 2010-03-03 11 -487 val_487 2010-03-03 11 -70 val_70 2010-03-03 11 -263 val_263 2010-03-03 11 -256 val_256 2010-03-03 11 -223 val_223 2010-03-03 11 -116 val_116 2010-03-03 11 -485 val_485 2010-03-03 11 -239 val_239 2010-03-03 11 -219 val_219 2010-03-03 11 -274 val_274 2010-03-03 11 -167 val_167 2010-03-03 11 -344 val_344 2010-03-03 11 -367 val_367 2010-03-03 11 -216 val_216 2010-03-03 11 -113 val_113 2010-03-03 11 -296 val_296 2010-03-03 11 -103 val_103 2010-03-03 11 -368 val_368 2010-03-03 11 -33 val_33 2010-03-03 11 -230 val_230 2010-03-03 11 -69 val_69 2010-03-03 11 -342 val_342 2010-03-03 11 -74 val_74 2010-03-03 11 -76 val_76 2010-03-03 11 -468 val_468 2010-03-03 11 -64 val_64 2010-03-03 11 +202 val_202 2010-03-03 11 +203 val_203 2010-03-03 11 +203 val_203 2010-03-03 11 +203 val_203 2010-03-03 11 +203 val_203 2010-03-03 11 +205 val_205 2010-03-03 11 +205 val_205 2010-03-03 11 +205 val_205 2010-03-03 11 +205 val_205 2010-03-03 11 +207 val_207 2010-03-03 11 +207 val_207 2010-03-03 11 +207 val_207 2010-03-03 11 +207 val_207 2010-03-03 11 +208 val_208 2010-03-03 11 +208 val_208 2010-03-03 11 +208 val_208 2010-03-03 11 +208 val_208 2010-03-03 11 +208 val_208 2010-03-03 11 +208 val_208 2010-03-03 11 209 val_209 2010-03-03 11 -30 val_30 2010-03-03 11 -453 val_453 2010-03-03 11 -138 val_138 2010-03-03 11 -228 val_228 2010-03-03 11 +209 val_209 2010-03-03 11 +209 val_209 2010-03-03 11 +209 val_209 2010-03-03 11 +213 val_213 2010-03-03 11 +213 val_213 2010-03-03 11 +213 val_213 2010-03-03 11 +213 val_213 2010-03-03 11 +214 val_214 2010-03-03 11 +214 val_214 2010-03-03 11 +216 val_216 2010-03-03 11 +216 val_216 2010-03-03 11 +216 val_216 2010-03-03 11 +216 val_216 2010-03-03 11 +217 val_217 2010-03-03 11 +217 val_217 2010-03-03 11 +217 val_217 2010-03-03 11 +217 val_217 2010-03-03 11 218 val_218 2010-03-03 11 -449 val_449 2010-03-03 11 -149 val_149 2010-03-03 11 -492 val_492 2010-03-03 11 +218 val_218 2010-03-03 11 +219 val_219 2010-03-03 11 +219 val_219 2010-03-03 11 +219 val_219 2010-03-03 11 +219 val_219 2010-03-03 11 +221 val_221 2010-03-03 11 +221 val_221 2010-03-03 11 +221 val_221 2010-03-03 11 +221 val_221 2010-03-03 11 +222 val_222 2010-03-03 11 +222 val_222 2010-03-03 11 223 val_223 2010-03-03 11 -41 val_41 2010-03-03 11 -76 val_76 2010-03-03 11 -78 val_78 2010-03-03 11 -458 val_458 2010-03-03 11 -489 val_489 2010-03-03 11 -119 val_119 2010-03-03 11 -430 val_430 2010-03-03 11 -321 val_321 2010-03-03 11 -42 val_42 2010-03-03 11 -195 val_195 2010-03-03 11 -160 val_160 2010-03-03 11 -498 val_498 2010-03-03 11 -322 val_322 2010-03-03 11 -472 val_472 2010-03-03 11 -143 val_143 2010-03-03 11 -233 val_233 2010-03-03 11 +223 val_223 2010-03-03 11 +223 val_223 2010-03-03 11 +223 val_223 2010-03-03 11 +224 val_224 2010-03-03 11 +224 val_224 2010-03-03 11 +224 val_224 2010-03-03 11 +224 val_224 2010-03-03 11 +226 val_226 2010-03-03 11 +226 val_226 2010-03-03 11 +228 val_228 2010-03-03 11 +228 val_228 2010-03-03 11 229 val_229 2010-03-03 11 -34 val_34 2010-03-03 11 -168 val_168 2010-03-03 11 -11 val_11 2010-03-03 11 -95 val_95 2010-03-03 11 -336 val_336 2010-03-03 11 -35 val_35 2010-03-03 11 -58 val_58 2010-03-03 11 -395 val_395 2010-03-03 11 -317 val_317 2010-03-03 11 -396 val_396 2010-03-03 11 -402 val_402 2010-03-03 11 -497 val_497 2010-03-03 11 -5 val_5 2010-03-03 11 -226 val_226 2010-03-03 11 -177 val_177 2010-03-03 11 -452 val_452 2010-03-03 11 +229 val_229 2010-03-03 11 +229 val_229 2010-03-03 11 +229 val_229 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +230 val_230 2010-03-03 11 +233 val_233 2010-03-03 11 +233 val_233 2010-03-03 11 +233 val_233 2010-03-03 11 +233 val_233 2010-03-03 11 +235 val_235 2010-03-03 11 +235 val_235 2010-03-03 11 +237 val_237 2010-03-03 11 +237 val_237 2010-03-03 11 +237 val_237 2010-03-03 11 +237 val_237 2010-03-03 11 +238 val_238 2010-03-03 11 +238 val_238 2010-03-03 11 +238 val_238 2010-03-03 11 +238 val_238 2010-03-03 11 +239 val_239 2010-03-03 11 +239 val_239 2010-03-03 11 +239 val_239 2010-03-03 11 +239 val_239 2010-03-03 11 +24 val_24 2010-03-03 11 +24 val_24 2010-03-03 11 +24 val_24 2010-03-03 11 +24 val_24 2010-03-03 11 +241 val_241 2010-03-03 11 +241 val_241 2010-03-03 11 242 val_242 2010-03-03 11 -401 val_401 2010-03-03 11 -331 val_331 2010-03-03 11 -272 val_272 2010-03-03 11 -392 val_392 2010-03-03 11 -369 val_369 2010-03-03 11 242 val_242 2010-03-03 11 -327 val_327 2010-03-03 11 -389 val_389 2010-03-03 11 -309 val_309 2010-03-03 11 -224 val_224 2010-03-03 11 -306 val_306 2010-03-03 11 +242 val_242 2010-03-03 11 +242 val_242 2010-03-03 11 +244 val_244 2010-03-03 11 +244 val_244 2010-03-03 11 +247 val_247 2010-03-03 11 +247 val_247 2010-03-03 11 +248 val_248 2010-03-03 11 +248 val_248 2010-03-03 11 +249 val_249 2010-03-03 11 +249 val_249 2010-03-03 11 +252 val_252 2010-03-03 11 +252 val_252 2010-03-03 11 +255 val_255 2010-03-03 11 +255 val_255 2010-03-03 11 +255 val_255 2010-03-03 11 +255 val_255 2010-03-03 11 +256 val_256 2010-03-03 11 +256 val_256 2010-03-03 11 +256 val_256 2010-03-03 11 +256 val_256 2010-03-03 11 +257 val_257 2010-03-03 11 +257 val_257 2010-03-03 11 +258 val_258 2010-03-03 11 +258 val_258 2010-03-03 11 +26 val_26 2010-03-03 11 +26 val_26 2010-03-03 11 +26 val_26 2010-03-03 11 +26 val_26 2010-03-03 11 +260 val_260 2010-03-03 11 +260 val_260 2010-03-03 11 +262 val_262 2010-03-03 11 +262 val_262 2010-03-03 11 +263 val_263 2010-03-03 11 +263 val_263 2010-03-03 11 +265 val_265 2010-03-03 11 +265 val_265 2010-03-03 11 +265 val_265 2010-03-03 11 +265 val_265 2010-03-03 11 +266 val_266 2010-03-03 11 +266 val_266 2010-03-03 11 +27 val_27 2010-03-03 11 +27 val_27 2010-03-03 11 +272 val_272 2010-03-03 11 +272 val_272 2010-03-03 11 +272 val_272 2010-03-03 11 +272 val_272 2010-03-03 11 +273 val_273 2010-03-03 11 273 val_273 2010-03-03 11 +273 val_273 2010-03-03 11 +273 val_273 2010-03-03 11 +273 val_273 2010-03-03 11 +273 val_273 2010-03-03 11 +274 val_274 2010-03-03 11 +274 val_274 2010-03-03 11 +275 val_275 2010-03-03 11 +275 val_275 2010-03-03 11 277 val_277 2010-03-03 11 -10 val_10 2010-03-03 11 -435 val_435 2010-03-03 11 -19 val_19 2010-03-03 11 -307 val_307 2010-03-03 11 -90 val_90 2010-03-03 11 -72 val_72 2010-03-03 11 -118 val_118 2010-03-03 11 -15 val_15 2010-03-03 11 -419 val_419 2010-03-03 11 -238 val_238 2010-03-03 11 -138 val_138 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +277 val_277 2010-03-03 11 +278 val_278 2010-03-03 11 +278 val_278 2010-03-03 11 +278 val_278 2010-03-03 11 +278 val_278 2010-03-03 11 +28 val_28 2010-03-03 11 +28 val_28 2010-03-03 11 +280 val_280 2010-03-03 11 +280 val_280 2010-03-03 11 +280 val_280 2010-03-03 11 +280 val_280 2010-03-03 11 +281 val_281 2010-03-03 11 +281 val_281 2010-03-03 11 +281 val_281 2010-03-03 11 +281 val_281 2010-03-03 11 282 val_282 2010-03-03 11 -395 val_395 2010-03-03 11 -134 val_134 2010-03-03 11 -118 val_118 2010-03-03 11 -179 val_179 2010-03-03 11 -364 val_364 2010-03-03 11 -87 val_87 2010-03-03 11 -490 val_490 2010-03-03 11 -0 val_0 2010-03-03 11 -77 val_77 2010-03-03 11 -85 val_85 2010-03-03 11 -413 val_413 2010-03-03 11 -169 val_169 2010-03-03 11 -316 val_316 2010-03-03 11 -369 val_369 2010-03-03 11 -137 val_137 2010-03-03 11 -470 val_470 2010-03-03 11 -409 val_409 2010-03-03 11 -318 val_318 2010-03-03 11 -318 val_318 2010-03-03 11 -187 val_187 2010-03-03 11 -197 val_197 2010-03-03 11 282 val_282 2010-03-03 11 -98 val_98 2010-03-03 11 -457 val_457 2010-03-03 11 -481 val_481 2010-03-03 11 -288 val_288 2010-03-03 11 -196 val_196 2010-03-03 11 -95 val_95 2010-03-03 11 -308 val_308 2010-03-03 11 -468 val_468 2010-03-03 11 -469 val_469 2010-03-03 11 -156 val_156 2010-03-03 11 -436 val_436 2010-03-03 11 -43 val_43 2010-03-03 11 -404 val_404 2010-03-03 11 -51 val_51 2010-03-03 11 -131 val_131 2010-03-03 11 -120 val_120 2010-03-03 11 -205 val_205 2010-03-03 11 -230 val_230 2010-03-03 11 -327 val_327 2010-03-03 11 -165 val_165 2010-03-03 11 -26 val_26 2010-03-03 11 -96 val_96 2010-03-03 11 -418 val_418 2010-03-03 11 -191 val_191 2010-03-03 11 -199 val_199 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -454 val_454 2010-03-03 11 -393 val_393 2010-03-03 11 -468 val_468 2010-03-03 11 -197 val_197 2010-03-03 11 -322 val_322 2010-03-03 11 -0 val_0 2010-03-03 11 -496 val_496 2010-03-03 11 -119 val_119 2010-03-03 11 -158 val_158 2010-03-03 11 -129 val_129 2010-03-03 11 -42 val_42 2010-03-03 11 -172 val_172 2010-03-03 11 -179 val_179 2010-03-03 11 -431 val_431 2010-03-03 11 -463 val_463 2010-03-03 11 -24 val_24 2010-03-03 11 -348 val_348 2010-03-03 11 -208 val_208 2010-03-03 11 -230 val_230 2010-03-03 11 -411 val_411 2010-03-03 11 -8 val_8 2010-03-03 11 -58 val_58 2010-03-03 11 -466 val_466 2010-03-03 11 -348 val_348 2010-03-03 11 -84 val_84 2010-03-03 11 -217 val_217 2010-03-03 11 -138 val_138 2010-03-03 11 -272 val_272 2010-03-03 11 -373 val_373 2010-03-03 11 -353 val_353 2010-03-03 11 -489 val_489 2010-03-03 11 -384 val_384 2010-03-03 11 -404 val_404 2010-03-03 11 -260 val_260 2010-03-03 11 -67 val_67 2010-03-03 11 -181 val_181 2010-03-03 11 -230 val_230 2010-03-03 11 -12 val_12 2010-03-03 11 +282 val_282 2010-03-03 11 +282 val_282 2010-03-03 11 +283 val_283 2010-03-03 11 +283 val_283 2010-03-03 11 284 val_284 2010-03-03 11 -180 val_180 2010-03-03 11 -333 val_333 2010-03-03 11 -83 val_83 2010-03-03 11 -241 val_241 2010-03-03 11 -137 val_137 2010-03-03 11 -275 val_275 2010-03-03 11 -311 val_311 2010-03-03 11 -332 val_332 2010-03-03 11 -318 val_318 2010-03-03 11 -65 val_65 2010-03-03 11 -221 val_221 2010-03-03 11 +284 val_284 2010-03-03 11 +285 val_285 2010-03-03 11 +285 val_285 2010-03-03 11 +286 val_286 2010-03-03 11 +286 val_286 2010-03-03 11 +287 val_287 2010-03-03 11 +287 val_287 2010-03-03 11 +288 val_288 2010-03-03 11 +288 val_288 2010-03-03 11 +288 val_288 2010-03-03 11 +288 val_288 2010-03-03 11 289 val_289 2010-03-03 11 -176 val_176 2010-03-03 11 -278 val_278 2010-03-03 11 -430 val_430 2010-03-03 11 -216 val_216 2010-03-03 11 -213 val_213 2010-03-03 11 -239 val_239 2010-03-03 11 -103 val_103 2010-03-03 11 -138 val_138 2010-03-03 11 -51 val_51 2010-03-03 11 -459 val_459 2010-03-03 11 -54 val_54 2010-03-03 11 -176 val_176 2010-03-03 11 -187 val_187 2010-03-03 11 -286 val_286 2010-03-03 11 -192 val_192 2010-03-03 11 -469 val_469 2010-03-03 11 -437 val_437 2010-03-03 11 -386 val_386 2010-03-03 11 -125 val_125 2010-03-03 11 -498 val_498 2010-03-03 11 -382 val_382 2010-03-03 11 -169 val_169 2010-03-03 11 -399 val_399 2010-03-03 11 -356 val_356 2010-03-03 11 -208 val_208 2010-03-03 11 -277 val_277 2010-03-03 11 -427 val_427 2010-03-03 11 -35 val_35 2010-03-03 11 -280 val_280 2010-03-03 11 -4 val_4 2010-03-03 11 -72 val_72 2010-03-03 11 -47 val_47 2010-03-03 11 -111 val_111 2010-03-03 11 -92 val_92 2010-03-03 11 -221 val_221 2010-03-03 11 -378 val_378 2010-03-03 11 -157 val_157 2010-03-03 11 -489 val_489 2010-03-03 11 -20 val_20 2010-03-03 11 -170 val_170 2010-03-03 11 -129 val_129 2010-03-03 11 -345 val_345 2010-03-03 11 -438 val_438 2010-03-03 11 -149 val_149 2010-03-03 11 -205 val_205 2010-03-03 11 +289 val_289 2010-03-03 11 +291 val_291 2010-03-03 11 +291 val_291 2010-03-03 11 +292 val_292 2010-03-03 11 +292 val_292 2010-03-03 11 +296 val_296 2010-03-03 11 +296 val_296 2010-03-03 11 +298 val_298 2010-03-03 11 +298 val_298 2010-03-03 11 +298 val_298 2010-03-03 11 +298 val_298 2010-03-03 11 +298 val_298 2010-03-03 11 +298 val_298 2010-03-03 11 +30 val_30 2010-03-03 11 +30 val_30 2010-03-03 11 302 val_302 2010-03-03 11 -57 val_57 2010-03-03 11 -316 val_316 2010-03-03 11 +302 val_302 2010-03-03 11 +305 val_305 2010-03-03 11 +305 val_305 2010-03-03 11 +306 val_306 2010-03-03 11 +306 val_306 2010-03-03 11 +307 val_307 2010-03-03 11 +307 val_307 2010-03-03 11 +307 val_307 2010-03-03 11 +307 val_307 2010-03-03 11 +308 val_308 2010-03-03 11 +308 val_308 2010-03-03 11 +309 val_309 2010-03-03 11 +309 val_309 2010-03-03 11 +309 val_309 2010-03-03 11 +309 val_309 2010-03-03 11 +310 val_310 2010-03-03 11 +310 val_310 2010-03-03 11 311 val_311 2010-03-03 11 -128 val_128 2010-03-03 11 -455 val_455 2010-03-03 11 -0 val_0 2010-03-03 11 -339 val_339 2010-03-03 11 -203 val_203 2010-03-03 11 -155 val_155 2010-03-03 11 -113 val_113 2010-03-03 11 -17 val_17 2010-03-03 11 -475 val_475 2010-03-03 11 -195 val_195 2010-03-03 11 -167 val_167 2010-03-03 11 +311 val_311 2010-03-03 11 +311 val_311 2010-03-03 11 +311 val_311 2010-03-03 11 +311 val_311 2010-03-03 11 +311 val_311 2010-03-03 11 +315 val_315 2010-03-03 11 +315 val_315 2010-03-03 11 +316 val_316 2010-03-03 11 +316 val_316 2010-03-03 11 +316 val_316 2010-03-03 11 +316 val_316 2010-03-03 11 +316 val_316 2010-03-03 11 +316 val_316 2010-03-03 11 +317 val_317 2010-03-03 11 +317 val_317 2010-03-03 11 +317 val_317 2010-03-03 11 +317 val_317 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +321 val_321 2010-03-03 11 +321 val_321 2010-03-03 11 +321 val_321 2010-03-03 11 +321 val_321 2010-03-03 11 +322 val_322 2010-03-03 11 +322 val_322 2010-03-03 11 +322 val_322 2010-03-03 11 +322 val_322 2010-03-03 11 +323 val_323 2010-03-03 11 +323 val_323 2010-03-03 11 325 val_325 2010-03-03 11 -367 val_367 2010-03-03 11 +325 val_325 2010-03-03 11 +325 val_325 2010-03-03 11 +325 val_325 2010-03-03 11 +327 val_327 2010-03-03 11 +327 val_327 2010-03-03 11 +327 val_327 2010-03-03 11 +327 val_327 2010-03-03 11 +327 val_327 2010-03-03 11 +327 val_327 2010-03-03 11 +33 val_33 2010-03-03 11 +33 val_33 2010-03-03 11 +331 val_331 2010-03-03 11 +331 val_331 2010-03-03 11 +331 val_331 2010-03-03 11 +331 val_331 2010-03-03 11 +332 val_332 2010-03-03 11 +332 val_332 2010-03-03 11 +333 val_333 2010-03-03 11 +333 val_333 2010-03-03 11 +333 val_333 2010-03-03 11 +333 val_333 2010-03-03 11 +335 val_335 2010-03-03 11 +335 val_335 2010-03-03 11 +336 val_336 2010-03-03 11 +336 val_336 2010-03-03 11 +338 val_338 2010-03-03 11 +338 val_338 2010-03-03 11 +339 val_339 2010-03-03 11 +339 val_339 2010-03-03 11 +34 val_34 2010-03-03 11 +34 val_34 2010-03-03 11 +341 val_341 2010-03-03 11 +341 val_341 2010-03-03 11 342 val_342 2010-03-03 11 -439 val_439 2010-03-03 11 -266 val_266 2010-03-03 11 +342 val_342 2010-03-03 11 +342 val_342 2010-03-03 11 +342 val_342 2010-03-03 11 +344 val_344 2010-03-03 11 +344 val_344 2010-03-03 11 +344 val_344 2010-03-03 11 +344 val_344 2010-03-03 11 +345 val_345 2010-03-03 11 +345 val_345 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +348 val_348 2010-03-03 11 +35 val_35 2010-03-03 11 +35 val_35 2010-03-03 11 +35 val_35 2010-03-03 11 +35 val_35 2010-03-03 11 +35 val_35 2010-03-03 11 +35 val_35 2010-03-03 11 +351 val_351 2010-03-03 11 +351 val_351 2010-03-03 11 +353 val_353 2010-03-03 11 +353 val_353 2010-03-03 11 +353 val_353 2010-03-03 11 +353 val_353 2010-03-03 11 +356 val_356 2010-03-03 11 +356 val_356 2010-03-03 11 +360 val_360 2010-03-03 11 +360 val_360 2010-03-03 11 +362 val_362 2010-03-03 11 +362 val_362 2010-03-03 11 +364 val_364 2010-03-03 11 +364 val_364 2010-03-03 11 365 val_365 2010-03-03 11 -309 val_309 2010-03-03 11 -397 val_397 2010-03-03 11 -377 val_377 2010-03-03 11 -162 val_162 2010-03-03 11 -489 val_489 2010-03-03 11 -417 val_417 2010-03-03 11 -247 val_247 2010-03-03 11 -396 val_396 2010-03-03 11 -399 val_399 2010-03-03 11 -174 val_174 2010-03-03 11 -208 val_208 2010-03-03 11 -466 val_466 2010-03-03 11 -199 val_199 2010-03-03 11 -207 val_207 2010-03-03 11 -494 val_494 2010-03-03 11 -413 val_413 2010-03-03 11 -174 val_174 2010-03-03 11 -482 val_482 2010-03-03 11 -237 val_237 2010-03-03 11 -394 val_394 2010-03-03 11 -459 val_459 2010-03-03 11 -446 val_446 2010-03-03 11 -338 val_338 2010-03-03 11 -193 val_193 2010-03-03 11 -153 val_153 2010-03-03 11 -287 val_287 2010-03-03 11 -219 val_219 2010-03-03 11 -292 val_292 2010-03-03 11 -252 val_252 2010-03-03 11 -430 val_430 2010-03-03 11 -417 val_417 2010-03-03 11 -166 val_166 2010-03-03 11 -403 val_403 2010-03-03 11 -82 val_82 2010-03-03 11 -15 val_15 2010-03-03 11 -209 val_209 2010-03-03 11 -277 val_277 2010-03-03 11 -281 val_281 2010-03-03 11 -327 val_327 2010-03-03 11 +365 val_365 2010-03-03 11 +366 val_366 2010-03-03 11 +366 val_366 2010-03-03 11 +367 val_367 2010-03-03 11 +367 val_367 2010-03-03 11 +367 val_367 2010-03-03 11 +367 val_367 2010-03-03 11 +368 val_368 2010-03-03 11 +368 val_368 2010-03-03 11 +369 val_369 2010-03-03 11 +369 val_369 2010-03-03 11 +369 val_369 2010-03-03 11 +369 val_369 2010-03-03 11 +369 val_369 2010-03-03 11 +369 val_369 2010-03-03 11 37 val_37 2010-03-03 11 -495 val_495 2010-03-03 11 -145 val_145 2010-03-03 11 -469 val_469 2010-03-03 11 -152 val_152 2010-03-03 11 -374 val_374 2010-03-03 11 -429 val_429 2010-03-03 11 -406 val_406 2010-03-03 11 -146 val_146 2010-03-03 11 -213 val_213 2010-03-03 11 -128 val_128 2010-03-03 11 -66 val_66 2010-03-03 11 -369 val_369 2010-03-03 11 -224 val_224 2010-03-03 11 -273 val_273 2010-03-03 11 -150 val_150 2010-03-03 11 -401 val_401 2010-03-03 11 -193 val_193 2010-03-03 11 -265 val_265 2010-03-03 11 -484 val_484 2010-03-03 11 -98 val_98 2010-03-03 11 -278 val_278 2010-03-03 11 -255 val_255 2010-03-03 11 -409 val_409 2010-03-03 11 -165 val_165 2010-03-03 11 -27 val_27 2010-03-03 11 -311 val_311 2010-03-03 11 -86 val_86 2010-03-03 11 -97 val_97 2010-03-03 11 -200 val_200 2010-03-03 11 -400 val_400 2010-03-03 11 -403 val_403 2010-03-03 11 -169 val_169 2010-03-03 11 -90 val_90 2010-03-03 11 -126 val_126 2010-03-03 11 -222 val_222 2010-03-03 11 -477 val_477 2010-03-03 11 -414 val_414 2010-03-03 11 -194 val_194 2010-03-03 11 -307 val_307 2010-03-03 11 -348 val_348 2010-03-03 11 -152 val_152 2010-03-03 11 -448 val_448 2010-03-03 11 37 val_37 2010-03-03 11 -28 val_28 2010-03-03 11 -84 val_84 2010-03-03 11 -315 val_315 2010-03-03 11 -469 val_469 2010-03-03 11 -97 val_97 2010-03-03 11 -344 val_344 2010-03-03 11 -281 val_281 2010-03-03 11 -183 val_183 2010-03-03 11 -273 val_273 2010-03-03 11 -18 val_18 2010-03-03 11 -167 val_167 2010-03-03 11 -348 val_348 2010-03-03 11 -285 val_285 2010-03-03 11 -186 val_186 2010-03-03 11 -362 val_362 2010-03-03 11 -458 val_458 2010-03-03 11 -146 val_146 2010-03-03 11 -498 val_498 2010-03-03 11 -341 val_341 2010-03-03 11 -9 val_9 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -492 val_492 2010-03-03 11 -462 val_462 2010-03-03 11 -18 val_18 2010-03-03 11 +37 val_37 2010-03-03 11 +37 val_37 2010-03-03 11 +373 val_373 2010-03-03 11 +373 val_373 2010-03-03 11 +374 val_374 2010-03-03 11 +374 val_374 2010-03-03 11 +375 val_375 2010-03-03 11 +375 val_375 2010-03-03 11 +377 val_377 2010-03-03 11 +377 val_377 2010-03-03 11 +378 val_378 2010-03-03 11 +378 val_378 2010-03-03 11 +379 val_379 2010-03-03 11 379 val_379 2010-03-03 11 +382 val_382 2010-03-03 11 +382 val_382 2010-03-03 11 +382 val_382 2010-03-03 11 +382 val_382 2010-03-03 11 384 val_384 2010-03-03 11 -67 val_67 2010-03-03 11 -134 val_134 2010-03-03 11 -26 val_26 2010-03-03 11 -256 val_256 2010-03-03 11 384 val_384 2010-03-03 11 -407 val_407 2010-03-03 11 -421 val_421 2010-03-03 11 +384 val_384 2010-03-03 11 +384 val_384 2010-03-03 11 +384 val_384 2010-03-03 11 +384 val_384 2010-03-03 11 +386 val_386 2010-03-03 11 +386 val_386 2010-03-03 11 +389 val_389 2010-03-03 11 +389 val_389 2010-03-03 11 +392 val_392 2010-03-03 11 +392 val_392 2010-03-03 11 +393 val_393 2010-03-03 11 +393 val_393 2010-03-03 11 +394 val_394 2010-03-03 11 +394 val_394 2010-03-03 11 +395 val_395 2010-03-03 11 +395 val_395 2010-03-03 11 +395 val_395 2010-03-03 11 +395 val_395 2010-03-03 11 +396 val_396 2010-03-03 11 +396 val_396 2010-03-03 11 +396 val_396 2010-03-03 11 +396 val_396 2010-03-03 11 +396 val_396 2010-03-03 11 +396 val_396 2010-03-03 11 +397 val_397 2010-03-03 11 +397 val_397 2010-03-03 11 +397 val_397 2010-03-03 11 +397 val_397 2010-03-03 11 +399 val_399 2010-03-03 11 +399 val_399 2010-03-03 11 +399 val_399 2010-03-03 11 +399 val_399 2010-03-03 11 +4 val_4 2010-03-03 11 +4 val_4 2010-03-03 11 +400 val_400 2010-03-03 11 +400 val_400 2010-03-03 11 401 val_401 2010-03-03 11 -375 val_375 2010-03-03 11 -454 val_454 2010-03-03 11 -189 val_189 2010-03-03 11 -175 val_175 2010-03-03 11 -133 val_133 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +401 val_401 2010-03-03 11 +402 val_402 2010-03-03 11 +402 val_402 2010-03-03 11 +403 val_403 2010-03-03 11 +403 val_403 2010-03-03 11 +403 val_403 2010-03-03 11 +403 val_403 2010-03-03 11 +403 val_403 2010-03-03 11 +403 val_403 2010-03-03 11 +404 val_404 2010-03-03 11 +404 val_404 2010-03-03 11 +404 val_404 2010-03-03 11 +404 val_404 2010-03-03 11 406 val_406 2010-03-03 11 -233 val_233 2010-03-03 11 -462 val_462 2010-03-03 11 -214 val_214 2010-03-03 11 -172 val_172 2010-03-03 11 -353 val_353 2010-03-03 11 -136 val_136 2010-03-03 11 -83 val_83 2010-03-03 11 -480 val_480 2010-03-03 11 -265 val_265 2010-03-03 11 -249 val_249 2010-03-03 11 -207 val_207 2010-03-03 11 -460 val_460 2010-03-03 11 -493 val_493 2010-03-03 11 -333 val_333 2010-03-03 11 -317 val_317 2010-03-03 11 -310 val_310 2010-03-03 11 -468 val_468 2010-03-03 11 -178 val_178 2010-03-03 11 -478 val_478 2010-03-03 11 -230 val_230 2010-03-03 11 -277 val_277 2010-03-03 11 -325 val_325 2010-03-03 11 -323 val_323 2010-03-03 11 -443 val_443 2010-03-03 11 -169 val_169 2010-03-03 11 -429 val_429 2010-03-03 11 -120 val_120 2010-03-03 11 -444 val_444 2010-03-03 11 -199 val_199 2010-03-03 11 -417 val_417 2010-03-03 11 -305 val_305 2010-03-03 11 -479 val_479 2010-03-03 11 -248 val_248 2010-03-03 11 -360 val_360 2010-03-03 11 -439 val_439 2010-03-03 11 -237 val_237 2010-03-03 11 -491 val_491 2010-03-03 11 -200 val_200 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +406 val_406 2010-03-03 11 +407 val_407 2010-03-03 11 +407 val_407 2010-03-03 11 +409 val_409 2010-03-03 11 +409 val_409 2010-03-03 11 +409 val_409 2010-03-03 11 +409 val_409 2010-03-03 11 +409 val_409 2010-03-03 11 +409 val_409 2010-03-03 11 +41 val_41 2010-03-03 11 +41 val_41 2010-03-03 11 +411 val_411 2010-03-03 11 +411 val_411 2010-03-03 11 +413 val_413 2010-03-03 11 +413 val_413 2010-03-03 11 +413 val_413 2010-03-03 11 +413 val_413 2010-03-03 11 414 val_414 2010-03-03 11 -119 val_119 2010-03-03 11 -438 val_438 2010-03-03 11 -163 val_163 2010-03-03 11 -70 val_70 2010-03-03 11 -104 val_104 2010-03-03 11 -255 val_255 2010-03-03 11 -351 val_351 2010-03-03 11 -24 val_24 2010-03-03 11 -291 val_291 2010-03-03 11 -480 val_480 2010-03-03 11 -397 val_397 2010-03-03 11 -70 val_70 2010-03-03 11 -5 val_5 2010-03-03 11 -382 val_382 2010-03-03 11 -187 val_187 2010-03-03 11 +414 val_414 2010-03-03 11 +414 val_414 2010-03-03 11 +414 val_414 2010-03-03 11 +417 val_417 2010-03-03 11 +417 val_417 2010-03-03 11 +417 val_417 2010-03-03 11 +417 val_417 2010-03-03 11 +417 val_417 2010-03-03 11 +417 val_417 2010-03-03 11 +418 val_418 2010-03-03 11 +418 val_418 2010-03-03 11 +419 val_419 2010-03-03 11 +419 val_419 2010-03-03 11 +42 val_42 2010-03-03 11 +42 val_42 2010-03-03 11 +42 val_42 2010-03-03 11 +42 val_42 2010-03-03 11 +421 val_421 2010-03-03 11 +421 val_421 2010-03-03 11 424 val_424 2010-03-03 11 -164 val_164 2010-03-03 11 -431 val_431 2010-03-03 11 -125 val_125 2010-03-03 11 -298 val_298 2010-03-03 11 -478 val_478 2010-03-03 11 -454 val_454 2010-03-03 11 -431 val_431 2010-03-03 11 -164 val_164 2010-03-03 11 -217 val_217 2010-03-03 11 -201 val_201 2010-03-03 11 -396 val_396 2010-03-03 11 -12 val_12 2010-03-03 11 424 val_424 2010-03-03 11 -348 val_348 2010-03-03 11 -262 val_262 2010-03-03 11 -203 val_203 2010-03-03 11 -90 val_90 2010-03-03 11 -258 val_258 2010-03-03 11 -114 val_114 2010-03-03 11 -401 val_401 2010-03-03 11 -406 val_406 2010-03-03 11 -190 val_190 2010-03-03 11 -409 val_409 2010-03-03 11 -406 val_406 2010-03-03 11 -257 val_257 2010-03-03 11 -105 val_105 2010-03-03 11 -53 val_53 2010-03-03 11 -483 val_483 2010-03-03 11 -403 val_403 2010-03-03 11 -175 val_175 2010-03-03 11 -366 val_366 2010-03-03 11 -466 val_466 2010-03-03 11 -104 val_104 2010-03-03 11 -335 val_335 2010-03-03 11 -321 val_321 2010-03-03 11 -193 val_193 2010-03-03 11 +424 val_424 2010-03-03 11 +424 val_424 2010-03-03 11 +427 val_427 2010-03-03 11 +427 val_427 2010-03-03 11 +429 val_429 2010-03-03 11 +429 val_429 2010-03-03 11 +429 val_429 2010-03-03 11 +429 val_429 2010-03-03 11 +43 val_43 2010-03-03 11 +43 val_43 2010-03-03 11 +430 val_430 2010-03-03 11 +430 val_430 2010-03-03 11 +430 val_430 2010-03-03 11 +430 val_430 2010-03-03 11 +430 val_430 2010-03-03 11 +430 val_430 2010-03-03 11 +431 val_431 2010-03-03 11 +431 val_431 2010-03-03 11 +431 val_431 2010-03-03 11 +431 val_431 2010-03-03 11 +431 val_431 2010-03-03 11 +431 val_431 2010-03-03 11 +432 val_432 2010-03-03 11 +432 val_432 2010-03-03 11 +435 val_435 2010-03-03 11 +435 val_435 2010-03-03 11 +436 val_436 2010-03-03 11 +436 val_436 2010-03-03 11 +437 val_437 2010-03-03 11 +437 val_437 2010-03-03 11 +438 val_438 2010-03-03 11 +438 val_438 2010-03-03 11 +438 val_438 2010-03-03 11 +438 val_438 2010-03-03 11 +438 val_438 2010-03-03 11 +438 val_438 2010-03-03 11 +439 val_439 2010-03-03 11 +439 val_439 2010-03-03 11 +439 val_439 2010-03-03 11 +439 val_439 2010-03-03 11 44 val_44 2010-03-03 11 -80 val_80 2010-03-03 11 -235 val_235 2010-03-03 11 -331 val_331 2010-03-03 11 -283 val_283 2010-03-03 11 -35 val_35 2010-03-03 11 -2 val_2 2010-03-03 11 -280 val_280 2010-03-03 11 +44 val_44 2010-03-03 11 +443 val_443 2010-03-03 11 +443 val_443 2010-03-03 11 +444 val_444 2010-03-03 11 +444 val_444 2010-03-03 11 +446 val_446 2010-03-03 11 +446 val_446 2010-03-03 11 +448 val_448 2010-03-03 11 +448 val_448 2010-03-03 11 +449 val_449 2010-03-03 11 +449 val_449 2010-03-03 11 +452 val_452 2010-03-03 11 +452 val_452 2010-03-03 11 +453 val_453 2010-03-03 11 +453 val_453 2010-03-03 11 +454 val_454 2010-03-03 11 +454 val_454 2010-03-03 11 +454 val_454 2010-03-03 11 +454 val_454 2010-03-03 11 +454 val_454 2010-03-03 11 +454 val_454 2010-03-03 11 +455 val_455 2010-03-03 11 +455 val_455 2010-03-03 11 +457 val_457 2010-03-03 11 +457 val_457 2010-03-03 11 +458 val_458 2010-03-03 11 +458 val_458 2010-03-03 11 +458 val_458 2010-03-03 11 +458 val_458 2010-03-03 11 +459 val_459 2010-03-03 11 +459 val_459 2010-03-03 11 +459 val_459 2010-03-03 11 +459 val_459 2010-03-03 11 +460 val_460 2010-03-03 11 +460 val_460 2010-03-03 11 +462 val_462 2010-03-03 11 +462 val_462 2010-03-03 11 +462 val_462 2010-03-03 11 +462 val_462 2010-03-03 11 463 val_463 2010-03-03 11 -469 val_469 2010-03-03 11 -229 val_229 2010-03-03 11 -316 val_316 2010-03-03 11 -202 val_202 2010-03-03 11 -432 val_432 2010-03-03 11 +463 val_463 2010-03-03 11 +463 val_463 2010-03-03 11 +463 val_463 2010-03-03 11 +466 val_466 2010-03-03 11 +466 val_466 2010-03-03 11 +466 val_466 2010-03-03 11 +466 val_466 2010-03-03 11 +466 val_466 2010-03-03 11 +466 val_466 2010-03-03 11 467 val_467 2010-03-03 11 -128 val_128 2010-03-03 11 -438 val_438 2010-03-03 11 -244 val_244 2010-03-03 11 -5 val_5 2010-03-03 11 -191 val_191 2010-03-03 11 -288 val_288 2010-03-03 11 -401 val_401 2010-03-03 11 +467 val_467 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +468 val_468 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +469 val_469 2010-03-03 11 +47 val_47 2010-03-03 11 +47 val_47 2010-03-03 11 +470 val_470 2010-03-03 11 +470 val_470 2010-03-03 11 +472 val_472 2010-03-03 11 +472 val_472 2010-03-03 11 +475 val_475 2010-03-03 11 +475 val_475 2010-03-03 11 +477 val_477 2010-03-03 11 +477 val_477 2010-03-03 11 +478 val_478 2010-03-03 11 +478 val_478 2010-03-03 11 +478 val_478 2010-03-03 11 +478 val_478 2010-03-03 11 +479 val_479 2010-03-03 11 +479 val_479 2010-03-03 11 480 val_480 2010-03-03 11 -487 val_487 2010-03-03 11 -70 val_70 2010-03-03 11 -263 val_263 2010-03-03 11 -256 val_256 2010-03-03 11 -223 val_223 2010-03-03 11 -116 val_116 2010-03-03 11 +480 val_480 2010-03-03 11 +480 val_480 2010-03-03 11 +480 val_480 2010-03-03 11 +480 val_480 2010-03-03 11 +480 val_480 2010-03-03 11 +481 val_481 2010-03-03 11 +481 val_481 2010-03-03 11 +482 val_482 2010-03-03 11 +482 val_482 2010-03-03 11 +483 val_483 2010-03-03 11 +483 val_483 2010-03-03 11 +484 val_484 2010-03-03 11 +484 val_484 2010-03-03 11 485 val_485 2010-03-03 11 -239 val_239 2010-03-03 11 -219 val_219 2010-03-03 11 -274 val_274 2010-03-03 11 -167 val_167 2010-03-03 11 -344 val_344 2010-03-03 11 -367 val_367 2010-03-03 11 -216 val_216 2010-03-03 11 -113 val_113 2010-03-03 11 -296 val_296 2010-03-03 11 -103 val_103 2010-03-03 11 -368 val_368 2010-03-03 11 -33 val_33 2010-03-03 11 -230 val_230 2010-03-03 11 -69 val_69 2010-03-03 11 -342 val_342 2010-03-03 11 -74 val_74 2010-03-03 11 -76 val_76 2010-03-03 11 -468 val_468 2010-03-03 11 -64 val_64 2010-03-03 11 -209 val_209 2010-03-03 11 -30 val_30 2010-03-03 11 -453 val_453 2010-03-03 11 -138 val_138 2010-03-03 11 -228 val_228 2010-03-03 11 -218 val_218 2010-03-03 11 -449 val_449 2010-03-03 11 -149 val_149 2010-03-03 11 -492 val_492 2010-03-03 11 -223 val_223 2010-03-03 11 -41 val_41 2010-03-03 11 -76 val_76 2010-03-03 11 -78 val_78 2010-03-03 11 -458 val_458 2010-03-03 11 +485 val_485 2010-03-03 11 +487 val_487 2010-03-03 11 +487 val_487 2010-03-03 11 489 val_489 2010-03-03 11 -119 val_119 2010-03-03 11 -430 val_430 2010-03-03 11 -321 val_321 2010-03-03 11 -42 val_42 2010-03-03 11 -195 val_195 2010-03-03 11 -160 val_160 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +489 val_489 2010-03-03 11 +490 val_490 2010-03-03 11 +490 val_490 2010-03-03 11 +491 val_491 2010-03-03 11 +491 val_491 2010-03-03 11 +492 val_492 2010-03-03 11 +492 val_492 2010-03-03 11 +492 val_492 2010-03-03 11 +492 val_492 2010-03-03 11 +493 val_493 2010-03-03 11 +493 val_493 2010-03-03 11 +494 val_494 2010-03-03 11 +494 val_494 2010-03-03 11 +495 val_495 2010-03-03 11 +495 val_495 2010-03-03 11 +496 val_496 2010-03-03 11 +496 val_496 2010-03-03 11 +497 val_497 2010-03-03 11 +497 val_497 2010-03-03 11 +498 val_498 2010-03-03 11 +498 val_498 2010-03-03 11 +498 val_498 2010-03-03 11 +498 val_498 2010-03-03 11 +498 val_498 2010-03-03 11 498 val_498 2010-03-03 11 -322 val_322 2010-03-03 11 -472 val_472 2010-03-03 11 -143 val_143 2010-03-03 11 -233 val_233 2010-03-03 11 -229 val_229 2010-03-03 11 -34 val_34 2010-03-03 11 -168 val_168 2010-03-03 11 -11 val_11 2010-03-03 11 -95 val_95 2010-03-03 11 -336 val_336 2010-03-03 11 -35 val_35 2010-03-03 11 -58 val_58 2010-03-03 11 -395 val_395 2010-03-03 11 -317 val_317 2010-03-03 11 -396 val_396 2010-03-03 11 -402 val_402 2010-03-03 11 -497 val_497 2010-03-03 11 5 val_5 2010-03-03 11 -226 val_226 2010-03-03 11 -177 val_177 2010-03-03 11 -452 val_452 2010-03-03 11 -242 val_242 2010-03-03 11 -401 val_401 2010-03-03 11 -331 val_331 2010-03-03 11 -272 val_272 2010-03-03 11 -392 val_392 2010-03-03 11 -369 val_369 2010-03-03 11 -242 val_242 2010-03-03 11 -327 val_327 2010-03-03 11 -389 val_389 2010-03-03 11 -309 val_309 2010-03-03 11 -224 val_224 2010-03-03 11 -306 val_306 2010-03-03 11 -273 val_273 2010-03-03 11 -277 val_277 2010-03-03 11 -10 val_10 2010-03-03 11 -435 val_435 2010-03-03 11 -19 val_19 2010-03-03 11 -307 val_307 2010-03-03 11 -90 val_90 2010-03-03 11 -72 val_72 2010-03-03 11 -118 val_118 2010-03-03 11 -15 val_15 2010-03-03 11 -419 val_419 2010-03-03 11 -238 val_238 2010-03-03 11 -138 val_138 2010-03-03 11 -282 val_282 2010-03-03 11 -395 val_395 2010-03-03 11 -134 val_134 2010-03-03 11 -118 val_118 2010-03-03 11 -179 val_179 2010-03-03 11 -364 val_364 2010-03-03 11 -87 val_87 2010-03-03 11 -490 val_490 2010-03-03 11 -0 val_0 2010-03-03 11 -77 val_77 2010-03-03 11 -85 val_85 2010-03-03 11 -413 val_413 2010-03-03 11 -169 val_169 2010-03-03 11 -316 val_316 2010-03-03 11 -369 val_369 2010-03-03 11 -137 val_137 2010-03-03 11 -470 val_470 2010-03-03 11 -409 val_409 2010-03-03 11 -318 val_318 2010-03-03 11 -318 val_318 2010-03-03 11 -187 val_187 2010-03-03 11 -197 val_197 2010-03-03 11 -282 val_282 2010-03-03 11 -98 val_98 2010-03-03 11 -457 val_457 2010-03-03 11 -481 val_481 2010-03-03 11 -288 val_288 2010-03-03 11 -196 val_196 2010-03-03 11 -95 val_95 2010-03-03 11 -308 val_308 2010-03-03 11 -468 val_468 2010-03-03 11 -469 val_469 2010-03-03 11 -156 val_156 2010-03-03 11 -436 val_436 2010-03-03 11 -43 val_43 2010-03-03 11 -404 val_404 2010-03-03 11 +5 val_5 2010-03-03 11 +5 val_5 2010-03-03 11 +5 val_5 2010-03-03 11 +5 val_5 2010-03-03 11 +5 val_5 2010-03-03 11 51 val_51 2010-03-03 11 -131 val_131 2010-03-03 11 -120 val_120 2010-03-03 11 -205 val_205 2010-03-03 11 -230 val_230 2010-03-03 11 -327 val_327 2010-03-03 11 -165 val_165 2010-03-03 11 -26 val_26 2010-03-03 11 -96 val_96 2010-03-03 11 -418 val_418 2010-03-03 11 -191 val_191 2010-03-03 11 -199 val_199 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -454 val_454 2010-03-03 11 -393 val_393 2010-03-03 11 -468 val_468 2010-03-03 11 -197 val_197 2010-03-03 11 -322 val_322 2010-03-03 11 -0 val_0 2010-03-03 11 -496 val_496 2010-03-03 11 -119 val_119 2010-03-03 11 -158 val_158 2010-03-03 11 -129 val_129 2010-03-03 11 -42 val_42 2010-03-03 11 -172 val_172 2010-03-03 11 -179 val_179 2010-03-03 11 -431 val_431 2010-03-03 11 -463 val_463 2010-03-03 11 -24 val_24 2010-03-03 11 -348 val_348 2010-03-03 11 -208 val_208 2010-03-03 11 -230 val_230 2010-03-03 11 -411 val_411 2010-03-03 11 -8 val_8 2010-03-03 11 -58 val_58 2010-03-03 11 -466 val_466 2010-03-03 11 -348 val_348 2010-03-03 11 -84 val_84 2010-03-03 11 -217 val_217 2010-03-03 11 -138 val_138 2010-03-03 11 -272 val_272 2010-03-03 11 -373 val_373 2010-03-03 11 -353 val_353 2010-03-03 11 -489 val_489 2010-03-03 11 -384 val_384 2010-03-03 11 -404 val_404 2010-03-03 11 -260 val_260 2010-03-03 11 -67 val_67 2010-03-03 11 -181 val_181 2010-03-03 11 -230 val_230 2010-03-03 11 -12 val_12 2010-03-03 11 -284 val_284 2010-03-03 11 -180 val_180 2010-03-03 11 -333 val_333 2010-03-03 11 -83 val_83 2010-03-03 11 -241 val_241 2010-03-03 11 -137 val_137 2010-03-03 11 -275 val_275 2010-03-03 11 -311 val_311 2010-03-03 11 -332 val_332 2010-03-03 11 -318 val_318 2010-03-03 11 -65 val_65 2010-03-03 11 -221 val_221 2010-03-03 11 -289 val_289 2010-03-03 11 -176 val_176 2010-03-03 11 -278 val_278 2010-03-03 11 -430 val_430 2010-03-03 11 -216 val_216 2010-03-03 11 -213 val_213 2010-03-03 11 -239 val_239 2010-03-03 11 -103 val_103 2010-03-03 11 -138 val_138 2010-03-03 11 51 val_51 2010-03-03 11 -459 val_459 2010-03-03 11 +51 val_51 2010-03-03 11 +51 val_51 2010-03-03 11 +53 val_53 2010-03-03 11 +53 val_53 2010-03-03 11 54 val_54 2010-03-03 11 -176 val_176 2010-03-03 11 -187 val_187 2010-03-03 11 -286 val_286 2010-03-03 11 -192 val_192 2010-03-03 11 -469 val_469 2010-03-03 11 -437 val_437 2010-03-03 11 -386 val_386 2010-03-03 11 -125 val_125 2010-03-03 11 -498 val_498 2010-03-03 11 -382 val_382 2010-03-03 11 -169 val_169 2010-03-03 11 -399 val_399 2010-03-03 11 -356 val_356 2010-03-03 11 -208 val_208 2010-03-03 11 -277 val_277 2010-03-03 11 -427 val_427 2010-03-03 11 -35 val_35 2010-03-03 11 -280 val_280 2010-03-03 11 -4 val_4 2010-03-03 11 +54 val_54 2010-03-03 11 +57 val_57 2010-03-03 11 +57 val_57 2010-03-03 11 +58 val_58 2010-03-03 11 +58 val_58 2010-03-03 11 +58 val_58 2010-03-03 11 +58 val_58 2010-03-03 11 +64 val_64 2010-03-03 11 +64 val_64 2010-03-03 11 +65 val_65 2010-03-03 11 +65 val_65 2010-03-03 11 +66 val_66 2010-03-03 11 +66 val_66 2010-03-03 11 +67 val_67 2010-03-03 11 +67 val_67 2010-03-03 11 +67 val_67 2010-03-03 11 +67 val_67 2010-03-03 11 +69 val_69 2010-03-03 11 +69 val_69 2010-03-03 11 +70 val_70 2010-03-03 11 +70 val_70 2010-03-03 11 +70 val_70 2010-03-03 11 +70 val_70 2010-03-03 11 +70 val_70 2010-03-03 11 +70 val_70 2010-03-03 11 72 val_72 2010-03-03 11 -47 val_47 2010-03-03 11 -111 val_111 2010-03-03 11 +72 val_72 2010-03-03 11 +72 val_72 2010-03-03 11 +72 val_72 2010-03-03 11 +74 val_74 2010-03-03 11 +74 val_74 2010-03-03 11 +76 val_76 2010-03-03 11 +76 val_76 2010-03-03 11 +76 val_76 2010-03-03 11 +76 val_76 2010-03-03 11 +77 val_77 2010-03-03 11 +77 val_77 2010-03-03 11 +78 val_78 2010-03-03 11 +78 val_78 2010-03-03 11 +8 val_8 2010-03-03 11 +8 val_8 2010-03-03 11 +80 val_80 2010-03-03 11 +80 val_80 2010-03-03 11 +82 val_82 2010-03-03 11 +82 val_82 2010-03-03 11 +83 val_83 2010-03-03 11 +83 val_83 2010-03-03 11 +83 val_83 2010-03-03 11 +83 val_83 2010-03-03 11 +84 val_84 2010-03-03 11 +84 val_84 2010-03-03 11 +84 val_84 2010-03-03 11 +84 val_84 2010-03-03 11 +85 val_85 2010-03-03 11 +85 val_85 2010-03-03 11 +86 val_86 2010-03-03 11 +86 val_86 2010-03-03 11 +87 val_87 2010-03-03 11 +87 val_87 2010-03-03 11 +9 val_9 2010-03-03 11 +9 val_9 2010-03-03 11 +90 val_90 2010-03-03 11 +90 val_90 2010-03-03 11 +90 val_90 2010-03-03 11 +90 val_90 2010-03-03 11 +90 val_90 2010-03-03 11 +90 val_90 2010-03-03 11 92 val_92 2010-03-03 11 -221 val_221 2010-03-03 11 -378 val_378 2010-03-03 11 -157 val_157 2010-03-03 11 -489 val_489 2010-03-03 11 -20 val_20 2010-03-03 11 -170 val_170 2010-03-03 11 -129 val_129 2010-03-03 11 -345 val_345 2010-03-03 11 -438 val_438 2010-03-03 11 -149 val_149 2010-03-03 11 -205 val_205 2010-03-03 11 -302 val_302 2010-03-03 11 -57 val_57 2010-03-03 11 -316 val_316 2010-03-03 11 -311 val_311 2010-03-03 11 -128 val_128 2010-03-03 11 -455 val_455 2010-03-03 11 -0 val_0 2010-03-03 11 -339 val_339 2010-03-03 11 -203 val_203 2010-03-03 11 -155 val_155 2010-03-03 11 -113 val_113 2010-03-03 11 -17 val_17 2010-03-03 11 -475 val_475 2010-03-03 11 -195 val_195 2010-03-03 11 -167 val_167 2010-03-03 11 -325 val_325 2010-03-03 11 -367 val_367 2010-03-03 11 -342 val_342 2010-03-03 11 -439 val_439 2010-03-03 11 -266 val_266 2010-03-03 11 -365 val_365 2010-03-03 11 -309 val_309 2010-03-03 11 -397 val_397 2010-03-03 11 -377 val_377 2010-03-03 11 -162 val_162 2010-03-03 11 -489 val_489 2010-03-03 11 -417 val_417 2010-03-03 11 -247 val_247 2010-03-03 11 -396 val_396 2010-03-03 11 -399 val_399 2010-03-03 11 -174 val_174 2010-03-03 11 -208 val_208 2010-03-03 11 -466 val_466 2010-03-03 11 -199 val_199 2010-03-03 11 -207 val_207 2010-03-03 11 -494 val_494 2010-03-03 11 -413 val_413 2010-03-03 11 -174 val_174 2010-03-03 11 -482 val_482 2010-03-03 11 -237 val_237 2010-03-03 11 -394 val_394 2010-03-03 11 -459 val_459 2010-03-03 11 -446 val_446 2010-03-03 11 -338 val_338 2010-03-03 11 -193 val_193 2010-03-03 11 -153 val_153 2010-03-03 11 -287 val_287 2010-03-03 11 -219 val_219 2010-03-03 11 -292 val_292 2010-03-03 11 -252 val_252 2010-03-03 11 -430 val_430 2010-03-03 11 -417 val_417 2010-03-03 11 -166 val_166 2010-03-03 11 -403 val_403 2010-03-03 11 -82 val_82 2010-03-03 11 -15 val_15 2010-03-03 11 -209 val_209 2010-03-03 11 -277 val_277 2010-03-03 11 -281 val_281 2010-03-03 11 -327 val_327 2010-03-03 11 -37 val_37 2010-03-03 11 -495 val_495 2010-03-03 11 -145 val_145 2010-03-03 11 -469 val_469 2010-03-03 11 -152 val_152 2010-03-03 11 -374 val_374 2010-03-03 11 -429 val_429 2010-03-03 11 -406 val_406 2010-03-03 11 -146 val_146 2010-03-03 11 -213 val_213 2010-03-03 11 -128 val_128 2010-03-03 11 -66 val_66 2010-03-03 11 -369 val_369 2010-03-03 11 -224 val_224 2010-03-03 11 -273 val_273 2010-03-03 11 -150 val_150 2010-03-03 11 -401 val_401 2010-03-03 11 -193 val_193 2010-03-03 11 -265 val_265 2010-03-03 11 -484 val_484 2010-03-03 11 +92 val_92 2010-03-03 11 +95 val_95 2010-03-03 11 +95 val_95 2010-03-03 11 +95 val_95 2010-03-03 11 +95 val_95 2010-03-03 11 +96 val_96 2010-03-03 11 +96 val_96 2010-03-03 11 +97 val_97 2010-03-03 11 +97 val_97 2010-03-03 11 +97 val_97 2010-03-03 11 +97 val_97 2010-03-03 11 +98 val_98 2010-03-03 11 +98 val_98 2010-03-03 11 +98 val_98 2010-03-03 11 98 val_98 2010-03-03 11 -278 val_278 2010-03-03 11 -255 val_255 2010-03-03 11 -409 val_409 2010-03-03 11 -165 val_165 2010-03-03 11 -27 val_27 2010-03-03 11 -311 val_311 2010-03-03 11 -86 val_86 2010-03-03 11 -238 val_238 2010-03-03 11 PREHOOK: query: select * from nzhang_part_n0 where ds = '2010-03-03' and hr = '12' PREHOOK: type: QUERY PREHOOK: Input: default@nzhang_part_n0 @@ -1074,1003 +1074,1003 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@nzhang_part_n0 POSTHOOK: Input: default@nzhang_part_n0@ds=2010-03-03/hr=12 #### A masked pattern was here #### -97 val_97 2010-03-03 12 -200 val_200 2010-03-03 12 -400 val_400 2010-03-03 12 -403 val_403 2010-03-03 12 -169 val_169 2010-03-03 12 -90 val_90 2010-03-03 12 -126 val_126 2010-03-03 12 -222 val_222 2010-03-03 12 -477 val_477 2010-03-03 12 -414 val_414 2010-03-03 12 -194 val_194 2010-03-03 12 -307 val_307 2010-03-03 12 -348 val_348 2010-03-03 12 -152 val_152 2010-03-03 12 -448 val_448 2010-03-03 12 -37 val_37 2010-03-03 12 -28 val_28 2010-03-03 12 -84 val_84 2010-03-03 12 -315 val_315 2010-03-03 12 -469 val_469 2010-03-03 12 -97 val_97 2010-03-03 12 -344 val_344 2010-03-03 12 -281 val_281 2010-03-03 12 -183 val_183 2010-03-03 12 -273 val_273 2010-03-03 12 -18 val_18 2010-03-03 12 -167 val_167 2010-03-03 12 -348 val_348 2010-03-03 12 -285 val_285 2010-03-03 12 -186 val_186 2010-03-03 12 -362 val_362 2010-03-03 12 -458 val_458 2010-03-03 12 -146 val_146 2010-03-03 12 -498 val_498 2010-03-03 12 -341 val_341 2010-03-03 12 -9 val_9 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -492 val_492 2010-03-03 12 -462 val_462 2010-03-03 12 -18 val_18 2010-03-03 12 -379 val_379 2010-03-03 12 -384 val_384 2010-03-03 12 -67 val_67 2010-03-03 12 -134 val_134 2010-03-03 12 -26 val_26 2010-03-03 12 -256 val_256 2010-03-03 12 -384 val_384 2010-03-03 12 -407 val_407 2010-03-03 12 -421 val_421 2010-03-03 12 -401 val_401 2010-03-03 12 -375 val_375 2010-03-03 12 -454 val_454 2010-03-03 12 -189 val_189 2010-03-03 12 -175 val_175 2010-03-03 12 -133 val_133 2010-03-03 12 -406 val_406 2010-03-03 12 -233 val_233 2010-03-03 12 -462 val_462 2010-03-03 12 -214 val_214 2010-03-03 12 -172 val_172 2010-03-03 12 -353 val_353 2010-03-03 12 -136 val_136 2010-03-03 12 -83 val_83 2010-03-03 12 -480 val_480 2010-03-03 12 -265 val_265 2010-03-03 12 -249 val_249 2010-03-03 12 -207 val_207 2010-03-03 12 -460 val_460 2010-03-03 12 -493 val_493 2010-03-03 12 -333 val_333 2010-03-03 12 -317 val_317 2010-03-03 12 -310 val_310 2010-03-03 12 -468 val_468 2010-03-03 12 -178 val_178 2010-03-03 12 -478 val_478 2010-03-03 12 -230 val_230 2010-03-03 12 -277 val_277 2010-03-03 12 -325 val_325 2010-03-03 12 -323 val_323 2010-03-03 12 -443 val_443 2010-03-03 12 -169 val_169 2010-03-03 12 -429 val_429 2010-03-03 12 -120 val_120 2010-03-03 12 -444 val_444 2010-03-03 12 -199 val_199 2010-03-03 12 -417 val_417 2010-03-03 12 -305 val_305 2010-03-03 12 -479 val_479 2010-03-03 12 -248 val_248 2010-03-03 12 -360 val_360 2010-03-03 12 -439 val_439 2010-03-03 12 -237 val_237 2010-03-03 12 -491 val_491 2010-03-03 12 -200 val_200 2010-03-03 12 -414 val_414 2010-03-03 12 -119 val_119 2010-03-03 12 -438 val_438 2010-03-03 12 -163 val_163 2010-03-03 12 -70 val_70 2010-03-03 12 -104 val_104 2010-03-03 12 -255 val_255 2010-03-03 12 -351 val_351 2010-03-03 12 -24 val_24 2010-03-03 12 -291 val_291 2010-03-03 12 -480 val_480 2010-03-03 12 -397 val_397 2010-03-03 12 -70 val_70 2010-03-03 12 -5 val_5 2010-03-03 12 -382 val_382 2010-03-03 12 -187 val_187 2010-03-03 12 -424 val_424 2010-03-03 12 -164 val_164 2010-03-03 12 -431 val_431 2010-03-03 12 -125 val_125 2010-03-03 12 -298 val_298 2010-03-03 12 -478 val_478 2010-03-03 12 -454 val_454 2010-03-03 12 -431 val_431 2010-03-03 12 -164 val_164 2010-03-03 12 -217 val_217 2010-03-03 12 -201 val_201 2010-03-03 12 -396 val_396 2010-03-03 12 -12 val_12 2010-03-03 12 -424 val_424 2010-03-03 12 -348 val_348 2010-03-03 12 -262 val_262 2010-03-03 12 -203 val_203 2010-03-03 12 -90 val_90 2010-03-03 12 -258 val_258 2010-03-03 12 -114 val_114 2010-03-03 12 -401 val_401 2010-03-03 12 -406 val_406 2010-03-03 12 -190 val_190 2010-03-03 12 -409 val_409 2010-03-03 12 -406 val_406 2010-03-03 12 -257 val_257 2010-03-03 12 -105 val_105 2010-03-03 12 -53 val_53 2010-03-03 12 -483 val_483 2010-03-03 12 -403 val_403 2010-03-03 12 -175 val_175 2010-03-03 12 -366 val_366 2010-03-03 12 -466 val_466 2010-03-03 12 -104 val_104 2010-03-03 12 -335 val_335 2010-03-03 12 -321 val_321 2010-03-03 12 -193 val_193 2010-03-03 12 -44 val_44 2010-03-03 12 -80 val_80 2010-03-03 12 -235 val_235 2010-03-03 12 -331 val_331 2010-03-03 12 -283 val_283 2010-03-03 12 -35 val_35 2010-03-03 12 -2 val_2 2010-03-03 12 -280 val_280 2010-03-03 12 -463 val_463 2010-03-03 12 -469 val_469 2010-03-03 12 -229 val_229 2010-03-03 12 -316 val_316 2010-03-03 12 -202 val_202 2010-03-03 12 -432 val_432 2010-03-03 12 -467 val_467 2010-03-03 12 -128 val_128 2010-03-03 12 -438 val_438 2010-03-03 12 -244 val_244 2010-03-03 12 -5 val_5 2010-03-03 12 -191 val_191 2010-03-03 12 -288 val_288 2010-03-03 12 -401 val_401 2010-03-03 12 -480 val_480 2010-03-03 12 -487 val_487 2010-03-03 12 -70 val_70 2010-03-03 12 -263 val_263 2010-03-03 12 -256 val_256 2010-03-03 12 -223 val_223 2010-03-03 12 -116 val_116 2010-03-03 12 -485 val_485 2010-03-03 12 -239 val_239 2010-03-03 12 -219 val_219 2010-03-03 12 -274 val_274 2010-03-03 12 -167 val_167 2010-03-03 12 -344 val_344 2010-03-03 12 -367 val_367 2010-03-03 12 -216 val_216 2010-03-03 12 -113 val_113 2010-03-03 12 -296 val_296 2010-03-03 12 -103 val_103 2010-03-03 12 -368 val_368 2010-03-03 12 -33 val_33 2010-03-03 12 -230 val_230 2010-03-03 12 -69 val_69 2010-03-03 12 -342 val_342 2010-03-03 12 -74 val_74 2010-03-03 12 -76 val_76 2010-03-03 12 -468 val_468 2010-03-03 12 -64 val_64 2010-03-03 12 -209 val_209 2010-03-03 12 -30 val_30 2010-03-03 12 -453 val_453 2010-03-03 12 -138 val_138 2010-03-03 12 -228 val_228 2010-03-03 12 -218 val_218 2010-03-03 12 -449 val_449 2010-03-03 12 -149 val_149 2010-03-03 12 -492 val_492 2010-03-03 12 -223 val_223 2010-03-03 12 -41 val_41 2010-03-03 12 -76 val_76 2010-03-03 12 -78 val_78 2010-03-03 12 -458 val_458 2010-03-03 12 -489 val_489 2010-03-03 12 -119 val_119 2010-03-03 12 -430 val_430 2010-03-03 12 -321 val_321 2010-03-03 12 -42 val_42 2010-03-03 12 -195 val_195 2010-03-03 12 -160 val_160 2010-03-03 12 -498 val_498 2010-03-03 12 -322 val_322 2010-03-03 12 -472 val_472 2010-03-03 12 -143 val_143 2010-03-03 12 -233 val_233 2010-03-03 12 -229 val_229 2010-03-03 12 -34 val_34 2010-03-03 12 -168 val_168 2010-03-03 12 -11 val_11 2010-03-03 12 -95 val_95 2010-03-03 12 -336 val_336 2010-03-03 12 -35 val_35 2010-03-03 12 -58 val_58 2010-03-03 12 -395 val_395 2010-03-03 12 -317 val_317 2010-03-03 12 -396 val_396 2010-03-03 12 -402 val_402 2010-03-03 12 -497 val_497 2010-03-03 12 -5 val_5 2010-03-03 12 -226 val_226 2010-03-03 12 -177 val_177 2010-03-03 12 -452 val_452 2010-03-03 12 -242 val_242 2010-03-03 12 -401 val_401 2010-03-03 12 -331 val_331 2010-03-03 12 -272 val_272 2010-03-03 12 -392 val_392 2010-03-03 12 -369 val_369 2010-03-03 12 -242 val_242 2010-03-03 12 -327 val_327 2010-03-03 12 -389 val_389 2010-03-03 12 -309 val_309 2010-03-03 12 -224 val_224 2010-03-03 12 -306 val_306 2010-03-03 12 -273 val_273 2010-03-03 12 -277 val_277 2010-03-03 12 -10 val_10 2010-03-03 12 -435 val_435 2010-03-03 12 -19 val_19 2010-03-03 12 -307 val_307 2010-03-03 12 -90 val_90 2010-03-03 12 -72 val_72 2010-03-03 12 -118 val_118 2010-03-03 12 -15 val_15 2010-03-03 12 -419 val_419 2010-03-03 12 -238 val_238 2010-03-03 12 -138 val_138 2010-03-03 12 -282 val_282 2010-03-03 12 -395 val_395 2010-03-03 12 -134 val_134 2010-03-03 12 -118 val_118 2010-03-03 12 -179 val_179 2010-03-03 12 -364 val_364 2010-03-03 12 -87 val_87 2010-03-03 12 -490 val_490 2010-03-03 12 0 val_0 2010-03-03 12 -77 val_77 2010-03-03 12 -85 val_85 2010-03-03 12 -413 val_413 2010-03-03 12 -169 val_169 2010-03-03 12 -316 val_316 2010-03-03 12 -369 val_369 2010-03-03 12 -137 val_137 2010-03-03 12 -470 val_470 2010-03-03 12 -409 val_409 2010-03-03 12 -318 val_318 2010-03-03 12 -318 val_318 2010-03-03 12 -187 val_187 2010-03-03 12 -197 val_197 2010-03-03 12 -282 val_282 2010-03-03 12 -98 val_98 2010-03-03 12 -457 val_457 2010-03-03 12 -481 val_481 2010-03-03 12 -288 val_288 2010-03-03 12 -196 val_196 2010-03-03 12 -95 val_95 2010-03-03 12 -308 val_308 2010-03-03 12 -468 val_468 2010-03-03 12 -469 val_469 2010-03-03 12 -156 val_156 2010-03-03 12 -436 val_436 2010-03-03 12 -43 val_43 2010-03-03 12 -404 val_404 2010-03-03 12 -51 val_51 2010-03-03 12 -131 val_131 2010-03-03 12 -120 val_120 2010-03-03 12 -205 val_205 2010-03-03 12 -230 val_230 2010-03-03 12 -327 val_327 2010-03-03 12 -165 val_165 2010-03-03 12 -26 val_26 2010-03-03 12 -96 val_96 2010-03-03 12 -418 val_418 2010-03-03 12 -191 val_191 2010-03-03 12 -199 val_199 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -454 val_454 2010-03-03 12 -393 val_393 2010-03-03 12 -468 val_468 2010-03-03 12 -197 val_197 2010-03-03 12 -322 val_322 2010-03-03 12 0 val_0 2010-03-03 12 -496 val_496 2010-03-03 12 -119 val_119 2010-03-03 12 -158 val_158 2010-03-03 12 -129 val_129 2010-03-03 12 -42 val_42 2010-03-03 12 -172 val_172 2010-03-03 12 -179 val_179 2010-03-03 12 -431 val_431 2010-03-03 12 -463 val_463 2010-03-03 12 -24 val_24 2010-03-03 12 -348 val_348 2010-03-03 12 -208 val_208 2010-03-03 12 -230 val_230 2010-03-03 12 -411 val_411 2010-03-03 12 -8 val_8 2010-03-03 12 -58 val_58 2010-03-03 12 -466 val_466 2010-03-03 12 -348 val_348 2010-03-03 12 -84 val_84 2010-03-03 12 -217 val_217 2010-03-03 12 -138 val_138 2010-03-03 12 -272 val_272 2010-03-03 12 -373 val_373 2010-03-03 12 -353 val_353 2010-03-03 12 -489 val_489 2010-03-03 12 -384 val_384 2010-03-03 12 -404 val_404 2010-03-03 12 -260 val_260 2010-03-03 12 -67 val_67 2010-03-03 12 -181 val_181 2010-03-03 12 -230 val_230 2010-03-03 12 -12 val_12 2010-03-03 12 -284 val_284 2010-03-03 12 -180 val_180 2010-03-03 12 -333 val_333 2010-03-03 12 -83 val_83 2010-03-03 12 -241 val_241 2010-03-03 12 -137 val_137 2010-03-03 12 -275 val_275 2010-03-03 12 -311 val_311 2010-03-03 12 -332 val_332 2010-03-03 12 -318 val_318 2010-03-03 12 -65 val_65 2010-03-03 12 -221 val_221 2010-03-03 12 -289 val_289 2010-03-03 12 -176 val_176 2010-03-03 12 -278 val_278 2010-03-03 12 -430 val_430 2010-03-03 12 -216 val_216 2010-03-03 12 -213 val_213 2010-03-03 12 -239 val_239 2010-03-03 12 +0 val_0 2010-03-03 12 +0 val_0 2010-03-03 12 +0 val_0 2010-03-03 12 +0 val_0 2010-03-03 12 +10 val_10 2010-03-03 12 +10 val_10 2010-03-03 12 +100 val_100 2010-03-03 12 +100 val_100 2010-03-03 12 +100 val_100 2010-03-03 12 +100 val_100 2010-03-03 12 103 val_103 2010-03-03 12 -138 val_138 2010-03-03 12 -51 val_51 2010-03-03 12 -459 val_459 2010-03-03 12 -54 val_54 2010-03-03 12 -176 val_176 2010-03-03 12 -187 val_187 2010-03-03 12 -286 val_286 2010-03-03 12 -192 val_192 2010-03-03 12 -469 val_469 2010-03-03 12 -437 val_437 2010-03-03 12 -386 val_386 2010-03-03 12 -125 val_125 2010-03-03 12 -498 val_498 2010-03-03 12 -382 val_382 2010-03-03 12 -169 val_169 2010-03-03 12 -399 val_399 2010-03-03 12 -356 val_356 2010-03-03 12 -208 val_208 2010-03-03 12 -277 val_277 2010-03-03 12 -427 val_427 2010-03-03 12 -35 val_35 2010-03-03 12 -280 val_280 2010-03-03 12 -4 val_4 2010-03-03 12 -72 val_72 2010-03-03 12 -47 val_47 2010-03-03 12 +103 val_103 2010-03-03 12 +103 val_103 2010-03-03 12 +103 val_103 2010-03-03 12 +104 val_104 2010-03-03 12 +104 val_104 2010-03-03 12 +104 val_104 2010-03-03 12 +104 val_104 2010-03-03 12 +105 val_105 2010-03-03 12 +105 val_105 2010-03-03 12 +11 val_11 2010-03-03 12 +11 val_11 2010-03-03 12 +111 val_111 2010-03-03 12 111 val_111 2010-03-03 12 -92 val_92 2010-03-03 12 -221 val_221 2010-03-03 12 -378 val_378 2010-03-03 12 -157 val_157 2010-03-03 12 -489 val_489 2010-03-03 12 -20 val_20 2010-03-03 12 -170 val_170 2010-03-03 12 -129 val_129 2010-03-03 12 -345 val_345 2010-03-03 12 -438 val_438 2010-03-03 12 -149 val_149 2010-03-03 12 -205 val_205 2010-03-03 12 -302 val_302 2010-03-03 12 -57 val_57 2010-03-03 12 -316 val_316 2010-03-03 12 -311 val_311 2010-03-03 12 -128 val_128 2010-03-03 12 -455 val_455 2010-03-03 12 -0 val_0 2010-03-03 12 -339 val_339 2010-03-03 12 -203 val_203 2010-03-03 12 -155 val_155 2010-03-03 12 113 val_113 2010-03-03 12 -17 val_17 2010-03-03 12 -475 val_475 2010-03-03 12 -195 val_195 2010-03-03 12 -167 val_167 2010-03-03 12 -325 val_325 2010-03-03 12 -367 val_367 2010-03-03 12 -342 val_342 2010-03-03 12 -439 val_439 2010-03-03 12 -266 val_266 2010-03-03 12 -365 val_365 2010-03-03 12 -309 val_309 2010-03-03 12 -397 val_397 2010-03-03 12 -377 val_377 2010-03-03 12 -162 val_162 2010-03-03 12 -489 val_489 2010-03-03 12 -417 val_417 2010-03-03 12 -247 val_247 2010-03-03 12 -396 val_396 2010-03-03 12 -399 val_399 2010-03-03 12 -174 val_174 2010-03-03 12 -208 val_208 2010-03-03 12 -466 val_466 2010-03-03 12 -199 val_199 2010-03-03 12 -207 val_207 2010-03-03 12 -494 val_494 2010-03-03 12 -413 val_413 2010-03-03 12 -174 val_174 2010-03-03 12 -482 val_482 2010-03-03 12 -237 val_237 2010-03-03 12 -394 val_394 2010-03-03 12 -459 val_459 2010-03-03 12 -446 val_446 2010-03-03 12 -338 val_338 2010-03-03 12 -193 val_193 2010-03-03 12 -153 val_153 2010-03-03 12 -287 val_287 2010-03-03 12 -219 val_219 2010-03-03 12 -292 val_292 2010-03-03 12 -252 val_252 2010-03-03 12 -430 val_430 2010-03-03 12 -417 val_417 2010-03-03 12 -166 val_166 2010-03-03 12 -403 val_403 2010-03-03 12 -82 val_82 2010-03-03 12 -15 val_15 2010-03-03 12 -209 val_209 2010-03-03 12 -277 val_277 2010-03-03 12 -281 val_281 2010-03-03 12 -327 val_327 2010-03-03 12 -37 val_37 2010-03-03 12 -495 val_495 2010-03-03 12 +113 val_113 2010-03-03 12 +113 val_113 2010-03-03 12 +113 val_113 2010-03-03 12 +114 val_114 2010-03-03 12 +114 val_114 2010-03-03 12 +116 val_116 2010-03-03 12 +116 val_116 2010-03-03 12 +118 val_118 2010-03-03 12 +118 val_118 2010-03-03 12 +118 val_118 2010-03-03 12 +118 val_118 2010-03-03 12 +119 val_119 2010-03-03 12 +119 val_119 2010-03-03 12 +119 val_119 2010-03-03 12 +119 val_119 2010-03-03 12 +119 val_119 2010-03-03 12 +119 val_119 2010-03-03 12 +12 val_12 2010-03-03 12 +12 val_12 2010-03-03 12 +12 val_12 2010-03-03 12 +12 val_12 2010-03-03 12 +120 val_120 2010-03-03 12 +120 val_120 2010-03-03 12 +120 val_120 2010-03-03 12 +120 val_120 2010-03-03 12 +125 val_125 2010-03-03 12 +125 val_125 2010-03-03 12 +125 val_125 2010-03-03 12 +125 val_125 2010-03-03 12 +126 val_126 2010-03-03 12 +126 val_126 2010-03-03 12 +128 val_128 2010-03-03 12 +128 val_128 2010-03-03 12 +128 val_128 2010-03-03 12 +128 val_128 2010-03-03 12 +128 val_128 2010-03-03 12 +128 val_128 2010-03-03 12 +129 val_129 2010-03-03 12 +129 val_129 2010-03-03 12 +129 val_129 2010-03-03 12 +129 val_129 2010-03-03 12 +131 val_131 2010-03-03 12 +131 val_131 2010-03-03 12 +133 val_133 2010-03-03 12 +133 val_133 2010-03-03 12 +134 val_134 2010-03-03 12 +134 val_134 2010-03-03 12 +134 val_134 2010-03-03 12 +134 val_134 2010-03-03 12 +136 val_136 2010-03-03 12 +136 val_136 2010-03-03 12 +137 val_137 2010-03-03 12 +137 val_137 2010-03-03 12 +137 val_137 2010-03-03 12 +137 val_137 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +138 val_138 2010-03-03 12 +143 val_143 2010-03-03 12 +143 val_143 2010-03-03 12 +145 val_145 2010-03-03 12 145 val_145 2010-03-03 12 -469 val_469 2010-03-03 12 -152 val_152 2010-03-03 12 -374 val_374 2010-03-03 12 -429 val_429 2010-03-03 12 -406 val_406 2010-03-03 12 146 val_146 2010-03-03 12 -213 val_213 2010-03-03 12 -128 val_128 2010-03-03 12 -66 val_66 2010-03-03 12 -369 val_369 2010-03-03 12 -224 val_224 2010-03-03 12 -273 val_273 2010-03-03 12 +146 val_146 2010-03-03 12 +146 val_146 2010-03-03 12 +146 val_146 2010-03-03 12 +149 val_149 2010-03-03 12 +149 val_149 2010-03-03 12 +149 val_149 2010-03-03 12 +149 val_149 2010-03-03 12 +15 val_15 2010-03-03 12 +15 val_15 2010-03-03 12 +15 val_15 2010-03-03 12 +15 val_15 2010-03-03 12 150 val_150 2010-03-03 12 -401 val_401 2010-03-03 12 -193 val_193 2010-03-03 12 -265 val_265 2010-03-03 12 -484 val_484 2010-03-03 12 -98 val_98 2010-03-03 12 -278 val_278 2010-03-03 12 -255 val_255 2010-03-03 12 -409 val_409 2010-03-03 12 +150 val_150 2010-03-03 12 +152 val_152 2010-03-03 12 +152 val_152 2010-03-03 12 +152 val_152 2010-03-03 12 +152 val_152 2010-03-03 12 +153 val_153 2010-03-03 12 +153 val_153 2010-03-03 12 +155 val_155 2010-03-03 12 +155 val_155 2010-03-03 12 +156 val_156 2010-03-03 12 +156 val_156 2010-03-03 12 +157 val_157 2010-03-03 12 +157 val_157 2010-03-03 12 +158 val_158 2010-03-03 12 +158 val_158 2010-03-03 12 +160 val_160 2010-03-03 12 +160 val_160 2010-03-03 12 +162 val_162 2010-03-03 12 +162 val_162 2010-03-03 12 +163 val_163 2010-03-03 12 +163 val_163 2010-03-03 12 +164 val_164 2010-03-03 12 +164 val_164 2010-03-03 12 +164 val_164 2010-03-03 12 +164 val_164 2010-03-03 12 165 val_165 2010-03-03 12 -27 val_27 2010-03-03 12 -311 val_311 2010-03-03 12 -86 val_86 2010-03-03 12 -238 val_238 2010-03-03 12 -200 val_200 2010-03-03 12 -400 val_400 2010-03-03 12 -403 val_403 2010-03-03 12 +165 val_165 2010-03-03 12 +165 val_165 2010-03-03 12 +165 val_165 2010-03-03 12 +166 val_166 2010-03-03 12 +166 val_166 2010-03-03 12 +167 val_167 2010-03-03 12 +167 val_167 2010-03-03 12 +167 val_167 2010-03-03 12 +167 val_167 2010-03-03 12 +167 val_167 2010-03-03 12 +167 val_167 2010-03-03 12 +168 val_168 2010-03-03 12 +168 val_168 2010-03-03 12 +169 val_169 2010-03-03 12 169 val_169 2010-03-03 12 -90 val_90 2010-03-03 12 -126 val_126 2010-03-03 12 -222 val_222 2010-03-03 12 -477 val_477 2010-03-03 12 -414 val_414 2010-03-03 12 -194 val_194 2010-03-03 12 -307 val_307 2010-03-03 12 -348 val_348 2010-03-03 12 -152 val_152 2010-03-03 12 -448 val_448 2010-03-03 12 -37 val_37 2010-03-03 12 -28 val_28 2010-03-03 12 -84 val_84 2010-03-03 12 -315 val_315 2010-03-03 12 -469 val_469 2010-03-03 12 -97 val_97 2010-03-03 12 -344 val_344 2010-03-03 12 -281 val_281 2010-03-03 12 -183 val_183 2010-03-03 12 -273 val_273 2010-03-03 12 +169 val_169 2010-03-03 12 +169 val_169 2010-03-03 12 +169 val_169 2010-03-03 12 +169 val_169 2010-03-03 12 +169 val_169 2010-03-03 12 +169 val_169 2010-03-03 12 +17 val_17 2010-03-03 12 +17 val_17 2010-03-03 12 +170 val_170 2010-03-03 12 +170 val_170 2010-03-03 12 +172 val_172 2010-03-03 12 +172 val_172 2010-03-03 12 +172 val_172 2010-03-03 12 +172 val_172 2010-03-03 12 +174 val_174 2010-03-03 12 +174 val_174 2010-03-03 12 +174 val_174 2010-03-03 12 +174 val_174 2010-03-03 12 +175 val_175 2010-03-03 12 +175 val_175 2010-03-03 12 +175 val_175 2010-03-03 12 +175 val_175 2010-03-03 12 +176 val_176 2010-03-03 12 +176 val_176 2010-03-03 12 +176 val_176 2010-03-03 12 +176 val_176 2010-03-03 12 +177 val_177 2010-03-03 12 +177 val_177 2010-03-03 12 +178 val_178 2010-03-03 12 +178 val_178 2010-03-03 12 +179 val_179 2010-03-03 12 +179 val_179 2010-03-03 12 +179 val_179 2010-03-03 12 +179 val_179 2010-03-03 12 18 val_18 2010-03-03 12 -167 val_167 2010-03-03 12 -348 val_348 2010-03-03 12 -285 val_285 2010-03-03 12 -186 val_186 2010-03-03 12 -362 val_362 2010-03-03 12 -458 val_458 2010-03-03 12 -146 val_146 2010-03-03 12 -498 val_498 2010-03-03 12 -341 val_341 2010-03-03 12 -9 val_9 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -492 val_492 2010-03-03 12 -462 val_462 2010-03-03 12 18 val_18 2010-03-03 12 -379 val_379 2010-03-03 12 -384 val_384 2010-03-03 12 -67 val_67 2010-03-03 12 -134 val_134 2010-03-03 12 -26 val_26 2010-03-03 12 -256 val_256 2010-03-03 12 -384 val_384 2010-03-03 12 -407 val_407 2010-03-03 12 -421 val_421 2010-03-03 12 -401 val_401 2010-03-03 12 -375 val_375 2010-03-03 12 -454 val_454 2010-03-03 12 +18 val_18 2010-03-03 12 +18 val_18 2010-03-03 12 +180 val_180 2010-03-03 12 +180 val_180 2010-03-03 12 +181 val_181 2010-03-03 12 +181 val_181 2010-03-03 12 +183 val_183 2010-03-03 12 +183 val_183 2010-03-03 12 +186 val_186 2010-03-03 12 +186 val_186 2010-03-03 12 +187 val_187 2010-03-03 12 +187 val_187 2010-03-03 12 +187 val_187 2010-03-03 12 +187 val_187 2010-03-03 12 +187 val_187 2010-03-03 12 +187 val_187 2010-03-03 12 189 val_189 2010-03-03 12 -175 val_175 2010-03-03 12 -133 val_133 2010-03-03 12 -406 val_406 2010-03-03 12 -233 val_233 2010-03-03 12 -462 val_462 2010-03-03 12 -214 val_214 2010-03-03 12 -172 val_172 2010-03-03 12 -353 val_353 2010-03-03 12 -136 val_136 2010-03-03 12 -83 val_83 2010-03-03 12 -480 val_480 2010-03-03 12 -265 val_265 2010-03-03 12 -249 val_249 2010-03-03 12 +189 val_189 2010-03-03 12 +19 val_19 2010-03-03 12 +19 val_19 2010-03-03 12 +190 val_190 2010-03-03 12 +190 val_190 2010-03-03 12 +191 val_191 2010-03-03 12 +191 val_191 2010-03-03 12 +191 val_191 2010-03-03 12 +191 val_191 2010-03-03 12 +192 val_192 2010-03-03 12 +192 val_192 2010-03-03 12 +193 val_193 2010-03-03 12 +193 val_193 2010-03-03 12 +193 val_193 2010-03-03 12 +193 val_193 2010-03-03 12 +193 val_193 2010-03-03 12 +193 val_193 2010-03-03 12 +194 val_194 2010-03-03 12 +194 val_194 2010-03-03 12 +195 val_195 2010-03-03 12 +195 val_195 2010-03-03 12 +195 val_195 2010-03-03 12 +195 val_195 2010-03-03 12 +196 val_196 2010-03-03 12 +196 val_196 2010-03-03 12 +197 val_197 2010-03-03 12 +197 val_197 2010-03-03 12 +197 val_197 2010-03-03 12 +197 val_197 2010-03-03 12 +199 val_199 2010-03-03 12 +199 val_199 2010-03-03 12 +199 val_199 2010-03-03 12 +199 val_199 2010-03-03 12 +199 val_199 2010-03-03 12 +199 val_199 2010-03-03 12 +2 val_2 2010-03-03 12 +2 val_2 2010-03-03 12 +20 val_20 2010-03-03 12 +20 val_20 2010-03-03 12 +200 val_200 2010-03-03 12 +200 val_200 2010-03-03 12 +200 val_200 2010-03-03 12 +200 val_200 2010-03-03 12 +201 val_201 2010-03-03 12 +201 val_201 2010-03-03 12 +202 val_202 2010-03-03 12 +202 val_202 2010-03-03 12 +203 val_203 2010-03-03 12 +203 val_203 2010-03-03 12 +203 val_203 2010-03-03 12 +203 val_203 2010-03-03 12 +205 val_205 2010-03-03 12 +205 val_205 2010-03-03 12 +205 val_205 2010-03-03 12 +205 val_205 2010-03-03 12 207 val_207 2010-03-03 12 -460 val_460 2010-03-03 12 -493 val_493 2010-03-03 12 -333 val_333 2010-03-03 12 -317 val_317 2010-03-03 12 -310 val_310 2010-03-03 12 -468 val_468 2010-03-03 12 -178 val_178 2010-03-03 12 -478 val_478 2010-03-03 12 +207 val_207 2010-03-03 12 +207 val_207 2010-03-03 12 +207 val_207 2010-03-03 12 +208 val_208 2010-03-03 12 +208 val_208 2010-03-03 12 +208 val_208 2010-03-03 12 +208 val_208 2010-03-03 12 +208 val_208 2010-03-03 12 +208 val_208 2010-03-03 12 +209 val_209 2010-03-03 12 +209 val_209 2010-03-03 12 +209 val_209 2010-03-03 12 +209 val_209 2010-03-03 12 +213 val_213 2010-03-03 12 +213 val_213 2010-03-03 12 +213 val_213 2010-03-03 12 +213 val_213 2010-03-03 12 +214 val_214 2010-03-03 12 +214 val_214 2010-03-03 12 +216 val_216 2010-03-03 12 +216 val_216 2010-03-03 12 +216 val_216 2010-03-03 12 +216 val_216 2010-03-03 12 +217 val_217 2010-03-03 12 +217 val_217 2010-03-03 12 +217 val_217 2010-03-03 12 +217 val_217 2010-03-03 12 +218 val_218 2010-03-03 12 +218 val_218 2010-03-03 12 +219 val_219 2010-03-03 12 +219 val_219 2010-03-03 12 +219 val_219 2010-03-03 12 +219 val_219 2010-03-03 12 +221 val_221 2010-03-03 12 +221 val_221 2010-03-03 12 +221 val_221 2010-03-03 12 +221 val_221 2010-03-03 12 +222 val_222 2010-03-03 12 +222 val_222 2010-03-03 12 +223 val_223 2010-03-03 12 +223 val_223 2010-03-03 12 +223 val_223 2010-03-03 12 +223 val_223 2010-03-03 12 +224 val_224 2010-03-03 12 +224 val_224 2010-03-03 12 +224 val_224 2010-03-03 12 +224 val_224 2010-03-03 12 +226 val_226 2010-03-03 12 +226 val_226 2010-03-03 12 +228 val_228 2010-03-03 12 +228 val_228 2010-03-03 12 +229 val_229 2010-03-03 12 +229 val_229 2010-03-03 12 +229 val_229 2010-03-03 12 +229 val_229 2010-03-03 12 230 val_230 2010-03-03 12 -277 val_277 2010-03-03 12 -325 val_325 2010-03-03 12 -323 val_323 2010-03-03 12 -443 val_443 2010-03-03 12 -169 val_169 2010-03-03 12 -429 val_429 2010-03-03 12 -120 val_120 2010-03-03 12 -444 val_444 2010-03-03 12 -199 val_199 2010-03-03 12 -417 val_417 2010-03-03 12 -305 val_305 2010-03-03 12 -479 val_479 2010-03-03 12 -248 val_248 2010-03-03 12 -360 val_360 2010-03-03 12 -439 val_439 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +230 val_230 2010-03-03 12 +233 val_233 2010-03-03 12 +233 val_233 2010-03-03 12 +233 val_233 2010-03-03 12 +233 val_233 2010-03-03 12 +235 val_235 2010-03-03 12 +235 val_235 2010-03-03 12 237 val_237 2010-03-03 12 -491 val_491 2010-03-03 12 -200 val_200 2010-03-03 12 -414 val_414 2010-03-03 12 -119 val_119 2010-03-03 12 -438 val_438 2010-03-03 12 -163 val_163 2010-03-03 12 -70 val_70 2010-03-03 12 -104 val_104 2010-03-03 12 -255 val_255 2010-03-03 12 -351 val_351 2010-03-03 12 +237 val_237 2010-03-03 12 +237 val_237 2010-03-03 12 +237 val_237 2010-03-03 12 +238 val_238 2010-03-03 12 +238 val_238 2010-03-03 12 +238 val_238 2010-03-03 12 +238 val_238 2010-03-03 12 +239 val_239 2010-03-03 12 +239 val_239 2010-03-03 12 +239 val_239 2010-03-03 12 +239 val_239 2010-03-03 12 24 val_24 2010-03-03 12 -291 val_291 2010-03-03 12 -480 val_480 2010-03-03 12 -397 val_397 2010-03-03 12 -70 val_70 2010-03-03 12 -5 val_5 2010-03-03 12 -382 val_382 2010-03-03 12 -187 val_187 2010-03-03 12 -424 val_424 2010-03-03 12 -164 val_164 2010-03-03 12 -431 val_431 2010-03-03 12 -125 val_125 2010-03-03 12 -298 val_298 2010-03-03 12 -478 val_478 2010-03-03 12 -454 val_454 2010-03-03 12 -431 val_431 2010-03-03 12 -164 val_164 2010-03-03 12 -217 val_217 2010-03-03 12 -201 val_201 2010-03-03 12 -396 val_396 2010-03-03 12 -12 val_12 2010-03-03 12 -424 val_424 2010-03-03 12 -348 val_348 2010-03-03 12 -262 val_262 2010-03-03 12 -203 val_203 2010-03-03 12 -90 val_90 2010-03-03 12 -258 val_258 2010-03-03 12 -114 val_114 2010-03-03 12 -401 val_401 2010-03-03 12 -406 val_406 2010-03-03 12 -190 val_190 2010-03-03 12 -409 val_409 2010-03-03 12 -406 val_406 2010-03-03 12 -257 val_257 2010-03-03 12 -105 val_105 2010-03-03 12 -53 val_53 2010-03-03 12 -483 val_483 2010-03-03 12 -403 val_403 2010-03-03 12 -175 val_175 2010-03-03 12 -366 val_366 2010-03-03 12 -466 val_466 2010-03-03 12 -104 val_104 2010-03-03 12 -335 val_335 2010-03-03 12 -321 val_321 2010-03-03 12 -193 val_193 2010-03-03 12 -44 val_44 2010-03-03 12 -80 val_80 2010-03-03 12 -235 val_235 2010-03-03 12 -331 val_331 2010-03-03 12 -283 val_283 2010-03-03 12 -35 val_35 2010-03-03 12 -2 val_2 2010-03-03 12 -280 val_280 2010-03-03 12 -463 val_463 2010-03-03 12 -469 val_469 2010-03-03 12 -229 val_229 2010-03-03 12 -316 val_316 2010-03-03 12 -202 val_202 2010-03-03 12 -432 val_432 2010-03-03 12 -467 val_467 2010-03-03 12 -128 val_128 2010-03-03 12 -438 val_438 2010-03-03 12 +24 val_24 2010-03-03 12 +24 val_24 2010-03-03 12 +24 val_24 2010-03-03 12 +241 val_241 2010-03-03 12 +241 val_241 2010-03-03 12 +242 val_242 2010-03-03 12 +242 val_242 2010-03-03 12 +242 val_242 2010-03-03 12 +242 val_242 2010-03-03 12 244 val_244 2010-03-03 12 -5 val_5 2010-03-03 12 -191 val_191 2010-03-03 12 -288 val_288 2010-03-03 12 -401 val_401 2010-03-03 12 -480 val_480 2010-03-03 12 -487 val_487 2010-03-03 12 -70 val_70 2010-03-03 12 -263 val_263 2010-03-03 12 +244 val_244 2010-03-03 12 +247 val_247 2010-03-03 12 +247 val_247 2010-03-03 12 +248 val_248 2010-03-03 12 +248 val_248 2010-03-03 12 +249 val_249 2010-03-03 12 +249 val_249 2010-03-03 12 +252 val_252 2010-03-03 12 +252 val_252 2010-03-03 12 +255 val_255 2010-03-03 12 +255 val_255 2010-03-03 12 +255 val_255 2010-03-03 12 +255 val_255 2010-03-03 12 256 val_256 2010-03-03 12 -223 val_223 2010-03-03 12 -116 val_116 2010-03-03 12 -485 val_485 2010-03-03 12 -239 val_239 2010-03-03 12 -219 val_219 2010-03-03 12 +256 val_256 2010-03-03 12 +256 val_256 2010-03-03 12 +256 val_256 2010-03-03 12 +257 val_257 2010-03-03 12 +257 val_257 2010-03-03 12 +258 val_258 2010-03-03 12 +258 val_258 2010-03-03 12 +26 val_26 2010-03-03 12 +26 val_26 2010-03-03 12 +26 val_26 2010-03-03 12 +26 val_26 2010-03-03 12 +260 val_260 2010-03-03 12 +260 val_260 2010-03-03 12 +262 val_262 2010-03-03 12 +262 val_262 2010-03-03 12 +263 val_263 2010-03-03 12 +263 val_263 2010-03-03 12 +265 val_265 2010-03-03 12 +265 val_265 2010-03-03 12 +265 val_265 2010-03-03 12 +265 val_265 2010-03-03 12 +266 val_266 2010-03-03 12 +266 val_266 2010-03-03 12 +27 val_27 2010-03-03 12 +27 val_27 2010-03-03 12 +272 val_272 2010-03-03 12 +272 val_272 2010-03-03 12 +272 val_272 2010-03-03 12 +272 val_272 2010-03-03 12 +273 val_273 2010-03-03 12 +273 val_273 2010-03-03 12 +273 val_273 2010-03-03 12 +273 val_273 2010-03-03 12 +273 val_273 2010-03-03 12 +273 val_273 2010-03-03 12 274 val_274 2010-03-03 12 -167 val_167 2010-03-03 12 -344 val_344 2010-03-03 12 -367 val_367 2010-03-03 12 -216 val_216 2010-03-03 12 -113 val_113 2010-03-03 12 +274 val_274 2010-03-03 12 +275 val_275 2010-03-03 12 +275 val_275 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +277 val_277 2010-03-03 12 +278 val_278 2010-03-03 12 +278 val_278 2010-03-03 12 +278 val_278 2010-03-03 12 +278 val_278 2010-03-03 12 +28 val_28 2010-03-03 12 +28 val_28 2010-03-03 12 +280 val_280 2010-03-03 12 +280 val_280 2010-03-03 12 +280 val_280 2010-03-03 12 +280 val_280 2010-03-03 12 +281 val_281 2010-03-03 12 +281 val_281 2010-03-03 12 +281 val_281 2010-03-03 12 +281 val_281 2010-03-03 12 +282 val_282 2010-03-03 12 +282 val_282 2010-03-03 12 +282 val_282 2010-03-03 12 +282 val_282 2010-03-03 12 +283 val_283 2010-03-03 12 +283 val_283 2010-03-03 12 +284 val_284 2010-03-03 12 +284 val_284 2010-03-03 12 +285 val_285 2010-03-03 12 +285 val_285 2010-03-03 12 +286 val_286 2010-03-03 12 +286 val_286 2010-03-03 12 +287 val_287 2010-03-03 12 +287 val_287 2010-03-03 12 +288 val_288 2010-03-03 12 +288 val_288 2010-03-03 12 +288 val_288 2010-03-03 12 +288 val_288 2010-03-03 12 +289 val_289 2010-03-03 12 +289 val_289 2010-03-03 12 +291 val_291 2010-03-03 12 +291 val_291 2010-03-03 12 +292 val_292 2010-03-03 12 +292 val_292 2010-03-03 12 296 val_296 2010-03-03 12 -103 val_103 2010-03-03 12 -368 val_368 2010-03-03 12 -33 val_33 2010-03-03 12 -230 val_230 2010-03-03 12 -69 val_69 2010-03-03 12 -342 val_342 2010-03-03 12 -74 val_74 2010-03-03 12 -76 val_76 2010-03-03 12 -468 val_468 2010-03-03 12 -64 val_64 2010-03-03 12 -209 val_209 2010-03-03 12 +296 val_296 2010-03-03 12 +298 val_298 2010-03-03 12 +298 val_298 2010-03-03 12 +298 val_298 2010-03-03 12 +298 val_298 2010-03-03 12 +298 val_298 2010-03-03 12 +298 val_298 2010-03-03 12 30 val_30 2010-03-03 12 -453 val_453 2010-03-03 12 -138 val_138 2010-03-03 12 -228 val_228 2010-03-03 12 -218 val_218 2010-03-03 12 -449 val_449 2010-03-03 12 -149 val_149 2010-03-03 12 -492 val_492 2010-03-03 12 -223 val_223 2010-03-03 12 -41 val_41 2010-03-03 12 -76 val_76 2010-03-03 12 -78 val_78 2010-03-03 12 -458 val_458 2010-03-03 12 -489 val_489 2010-03-03 12 -119 val_119 2010-03-03 12 -430 val_430 2010-03-03 12 +30 val_30 2010-03-03 12 +302 val_302 2010-03-03 12 +302 val_302 2010-03-03 12 +305 val_305 2010-03-03 12 +305 val_305 2010-03-03 12 +306 val_306 2010-03-03 12 +306 val_306 2010-03-03 12 +307 val_307 2010-03-03 12 +307 val_307 2010-03-03 12 +307 val_307 2010-03-03 12 +307 val_307 2010-03-03 12 +308 val_308 2010-03-03 12 +308 val_308 2010-03-03 12 +309 val_309 2010-03-03 12 +309 val_309 2010-03-03 12 +309 val_309 2010-03-03 12 +309 val_309 2010-03-03 12 +310 val_310 2010-03-03 12 +310 val_310 2010-03-03 12 +311 val_311 2010-03-03 12 +311 val_311 2010-03-03 12 +311 val_311 2010-03-03 12 +311 val_311 2010-03-03 12 +311 val_311 2010-03-03 12 +311 val_311 2010-03-03 12 +315 val_315 2010-03-03 12 +315 val_315 2010-03-03 12 +316 val_316 2010-03-03 12 +316 val_316 2010-03-03 12 +316 val_316 2010-03-03 12 +316 val_316 2010-03-03 12 +316 val_316 2010-03-03 12 +316 val_316 2010-03-03 12 +317 val_317 2010-03-03 12 +317 val_317 2010-03-03 12 +317 val_317 2010-03-03 12 +317 val_317 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +321 val_321 2010-03-03 12 +321 val_321 2010-03-03 12 +321 val_321 2010-03-03 12 321 val_321 2010-03-03 12 -42 val_42 2010-03-03 12 -195 val_195 2010-03-03 12 -160 val_160 2010-03-03 12 -498 val_498 2010-03-03 12 322 val_322 2010-03-03 12 -472 val_472 2010-03-03 12 -143 val_143 2010-03-03 12 -233 val_233 2010-03-03 12 -229 val_229 2010-03-03 12 -34 val_34 2010-03-03 12 -168 val_168 2010-03-03 12 -11 val_11 2010-03-03 12 -95 val_95 2010-03-03 12 +322 val_322 2010-03-03 12 +322 val_322 2010-03-03 12 +322 val_322 2010-03-03 12 +323 val_323 2010-03-03 12 +323 val_323 2010-03-03 12 +325 val_325 2010-03-03 12 +325 val_325 2010-03-03 12 +325 val_325 2010-03-03 12 +325 val_325 2010-03-03 12 +327 val_327 2010-03-03 12 +327 val_327 2010-03-03 12 +327 val_327 2010-03-03 12 +327 val_327 2010-03-03 12 +327 val_327 2010-03-03 12 +327 val_327 2010-03-03 12 +33 val_33 2010-03-03 12 +33 val_33 2010-03-03 12 +331 val_331 2010-03-03 12 +331 val_331 2010-03-03 12 +331 val_331 2010-03-03 12 +331 val_331 2010-03-03 12 +332 val_332 2010-03-03 12 +332 val_332 2010-03-03 12 +333 val_333 2010-03-03 12 +333 val_333 2010-03-03 12 +333 val_333 2010-03-03 12 +333 val_333 2010-03-03 12 +335 val_335 2010-03-03 12 +335 val_335 2010-03-03 12 +336 val_336 2010-03-03 12 336 val_336 2010-03-03 12 +338 val_338 2010-03-03 12 +338 val_338 2010-03-03 12 +339 val_339 2010-03-03 12 +339 val_339 2010-03-03 12 +34 val_34 2010-03-03 12 +34 val_34 2010-03-03 12 +341 val_341 2010-03-03 12 +341 val_341 2010-03-03 12 +342 val_342 2010-03-03 12 +342 val_342 2010-03-03 12 +342 val_342 2010-03-03 12 +342 val_342 2010-03-03 12 +344 val_344 2010-03-03 12 +344 val_344 2010-03-03 12 +344 val_344 2010-03-03 12 +344 val_344 2010-03-03 12 +345 val_345 2010-03-03 12 +345 val_345 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 +348 val_348 2010-03-03 12 35 val_35 2010-03-03 12 -58 val_58 2010-03-03 12 -395 val_395 2010-03-03 12 -317 val_317 2010-03-03 12 -396 val_396 2010-03-03 12 -402 val_402 2010-03-03 12 -497 val_497 2010-03-03 12 -5 val_5 2010-03-03 12 -226 val_226 2010-03-03 12 -177 val_177 2010-03-03 12 -452 val_452 2010-03-03 12 -242 val_242 2010-03-03 12 -401 val_401 2010-03-03 12 -331 val_331 2010-03-03 12 -272 val_272 2010-03-03 12 -392 val_392 2010-03-03 12 +35 val_35 2010-03-03 12 +35 val_35 2010-03-03 12 +35 val_35 2010-03-03 12 +35 val_35 2010-03-03 12 +35 val_35 2010-03-03 12 +351 val_351 2010-03-03 12 +351 val_351 2010-03-03 12 +353 val_353 2010-03-03 12 +353 val_353 2010-03-03 12 +353 val_353 2010-03-03 12 +353 val_353 2010-03-03 12 +356 val_356 2010-03-03 12 +356 val_356 2010-03-03 12 +360 val_360 2010-03-03 12 +360 val_360 2010-03-03 12 +362 val_362 2010-03-03 12 +362 val_362 2010-03-03 12 +364 val_364 2010-03-03 12 +364 val_364 2010-03-03 12 +365 val_365 2010-03-03 12 +365 val_365 2010-03-03 12 +366 val_366 2010-03-03 12 +366 val_366 2010-03-03 12 +367 val_367 2010-03-03 12 +367 val_367 2010-03-03 12 +367 val_367 2010-03-03 12 +367 val_367 2010-03-03 12 +368 val_368 2010-03-03 12 +368 val_368 2010-03-03 12 369 val_369 2010-03-03 12 -242 val_242 2010-03-03 12 -327 val_327 2010-03-03 12 +369 val_369 2010-03-03 12 +369 val_369 2010-03-03 12 +369 val_369 2010-03-03 12 +369 val_369 2010-03-03 12 +369 val_369 2010-03-03 12 +37 val_37 2010-03-03 12 +37 val_37 2010-03-03 12 +37 val_37 2010-03-03 12 +37 val_37 2010-03-03 12 +373 val_373 2010-03-03 12 +373 val_373 2010-03-03 12 +374 val_374 2010-03-03 12 +374 val_374 2010-03-03 12 +375 val_375 2010-03-03 12 +375 val_375 2010-03-03 12 +377 val_377 2010-03-03 12 +377 val_377 2010-03-03 12 +378 val_378 2010-03-03 12 +378 val_378 2010-03-03 12 +379 val_379 2010-03-03 12 +379 val_379 2010-03-03 12 +382 val_382 2010-03-03 12 +382 val_382 2010-03-03 12 +382 val_382 2010-03-03 12 +382 val_382 2010-03-03 12 +384 val_384 2010-03-03 12 +384 val_384 2010-03-03 12 +384 val_384 2010-03-03 12 +384 val_384 2010-03-03 12 +384 val_384 2010-03-03 12 +384 val_384 2010-03-03 12 +386 val_386 2010-03-03 12 +386 val_386 2010-03-03 12 389 val_389 2010-03-03 12 -309 val_309 2010-03-03 12 -224 val_224 2010-03-03 12 -306 val_306 2010-03-03 12 -273 val_273 2010-03-03 12 -277 val_277 2010-03-03 12 -10 val_10 2010-03-03 12 -435 val_435 2010-03-03 12 -19 val_19 2010-03-03 12 -307 val_307 2010-03-03 12 -90 val_90 2010-03-03 12 -72 val_72 2010-03-03 12 -118 val_118 2010-03-03 12 -15 val_15 2010-03-03 12 -419 val_419 2010-03-03 12 -238 val_238 2010-03-03 12 -138 val_138 2010-03-03 12 -282 val_282 2010-03-03 12 +389 val_389 2010-03-03 12 +392 val_392 2010-03-03 12 +392 val_392 2010-03-03 12 +393 val_393 2010-03-03 12 +393 val_393 2010-03-03 12 +394 val_394 2010-03-03 12 +394 val_394 2010-03-03 12 395 val_395 2010-03-03 12 -134 val_134 2010-03-03 12 -118 val_118 2010-03-03 12 -179 val_179 2010-03-03 12 -364 val_364 2010-03-03 12 -87 val_87 2010-03-03 12 -490 val_490 2010-03-03 12 -0 val_0 2010-03-03 12 -77 val_77 2010-03-03 12 -85 val_85 2010-03-03 12 -413 val_413 2010-03-03 12 -169 val_169 2010-03-03 12 -316 val_316 2010-03-03 12 -369 val_369 2010-03-03 12 -137 val_137 2010-03-03 12 -470 val_470 2010-03-03 12 -409 val_409 2010-03-03 12 -318 val_318 2010-03-03 12 -318 val_318 2010-03-03 12 -187 val_187 2010-03-03 12 -197 val_197 2010-03-03 12 -282 val_282 2010-03-03 12 -98 val_98 2010-03-03 12 -457 val_457 2010-03-03 12 -481 val_481 2010-03-03 12 -288 val_288 2010-03-03 12 -196 val_196 2010-03-03 12 -95 val_95 2010-03-03 12 -308 val_308 2010-03-03 12 -468 val_468 2010-03-03 12 -469 val_469 2010-03-03 12 -156 val_156 2010-03-03 12 -436 val_436 2010-03-03 12 -43 val_43 2010-03-03 12 +395 val_395 2010-03-03 12 +395 val_395 2010-03-03 12 +395 val_395 2010-03-03 12 +396 val_396 2010-03-03 12 +396 val_396 2010-03-03 12 +396 val_396 2010-03-03 12 +396 val_396 2010-03-03 12 +396 val_396 2010-03-03 12 +396 val_396 2010-03-03 12 +397 val_397 2010-03-03 12 +397 val_397 2010-03-03 12 +397 val_397 2010-03-03 12 +397 val_397 2010-03-03 12 +399 val_399 2010-03-03 12 +399 val_399 2010-03-03 12 +399 val_399 2010-03-03 12 +399 val_399 2010-03-03 12 +4 val_4 2010-03-03 12 +4 val_4 2010-03-03 12 +400 val_400 2010-03-03 12 +400 val_400 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +401 val_401 2010-03-03 12 +402 val_402 2010-03-03 12 +402 val_402 2010-03-03 12 +403 val_403 2010-03-03 12 +403 val_403 2010-03-03 12 +403 val_403 2010-03-03 12 +403 val_403 2010-03-03 12 +403 val_403 2010-03-03 12 +403 val_403 2010-03-03 12 404 val_404 2010-03-03 12 -51 val_51 2010-03-03 12 -131 val_131 2010-03-03 12 -120 val_120 2010-03-03 12 -205 val_205 2010-03-03 12 -230 val_230 2010-03-03 12 -327 val_327 2010-03-03 12 -165 val_165 2010-03-03 12 -26 val_26 2010-03-03 12 -96 val_96 2010-03-03 12 +404 val_404 2010-03-03 12 +404 val_404 2010-03-03 12 +404 val_404 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +406 val_406 2010-03-03 12 +407 val_407 2010-03-03 12 +407 val_407 2010-03-03 12 +409 val_409 2010-03-03 12 +409 val_409 2010-03-03 12 +409 val_409 2010-03-03 12 +409 val_409 2010-03-03 12 +409 val_409 2010-03-03 12 +409 val_409 2010-03-03 12 +41 val_41 2010-03-03 12 +41 val_41 2010-03-03 12 +411 val_411 2010-03-03 12 +411 val_411 2010-03-03 12 +413 val_413 2010-03-03 12 +413 val_413 2010-03-03 12 +413 val_413 2010-03-03 12 +413 val_413 2010-03-03 12 +414 val_414 2010-03-03 12 +414 val_414 2010-03-03 12 +414 val_414 2010-03-03 12 +414 val_414 2010-03-03 12 +417 val_417 2010-03-03 12 +417 val_417 2010-03-03 12 +417 val_417 2010-03-03 12 +417 val_417 2010-03-03 12 +417 val_417 2010-03-03 12 +417 val_417 2010-03-03 12 418 val_418 2010-03-03 12 -191 val_191 2010-03-03 12 -199 val_199 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -454 val_454 2010-03-03 12 -393 val_393 2010-03-03 12 -468 val_468 2010-03-03 12 -197 val_197 2010-03-03 12 -322 val_322 2010-03-03 12 -0 val_0 2010-03-03 12 -496 val_496 2010-03-03 12 -119 val_119 2010-03-03 12 -158 val_158 2010-03-03 12 -129 val_129 2010-03-03 12 +418 val_418 2010-03-03 12 +419 val_419 2010-03-03 12 +419 val_419 2010-03-03 12 42 val_42 2010-03-03 12 -172 val_172 2010-03-03 12 -179 val_179 2010-03-03 12 +42 val_42 2010-03-03 12 +42 val_42 2010-03-03 12 +42 val_42 2010-03-03 12 +421 val_421 2010-03-03 12 +421 val_421 2010-03-03 12 +424 val_424 2010-03-03 12 +424 val_424 2010-03-03 12 +424 val_424 2010-03-03 12 +424 val_424 2010-03-03 12 +427 val_427 2010-03-03 12 +427 val_427 2010-03-03 12 +429 val_429 2010-03-03 12 +429 val_429 2010-03-03 12 +429 val_429 2010-03-03 12 +429 val_429 2010-03-03 12 +43 val_43 2010-03-03 12 +43 val_43 2010-03-03 12 +430 val_430 2010-03-03 12 +430 val_430 2010-03-03 12 +430 val_430 2010-03-03 12 +430 val_430 2010-03-03 12 +430 val_430 2010-03-03 12 +430 val_430 2010-03-03 12 +431 val_431 2010-03-03 12 +431 val_431 2010-03-03 12 +431 val_431 2010-03-03 12 +431 val_431 2010-03-03 12 +431 val_431 2010-03-03 12 431 val_431 2010-03-03 12 +432 val_432 2010-03-03 12 +432 val_432 2010-03-03 12 +435 val_435 2010-03-03 12 +435 val_435 2010-03-03 12 +436 val_436 2010-03-03 12 +436 val_436 2010-03-03 12 +437 val_437 2010-03-03 12 +437 val_437 2010-03-03 12 +438 val_438 2010-03-03 12 +438 val_438 2010-03-03 12 +438 val_438 2010-03-03 12 +438 val_438 2010-03-03 12 +438 val_438 2010-03-03 12 +438 val_438 2010-03-03 12 +439 val_439 2010-03-03 12 +439 val_439 2010-03-03 12 +439 val_439 2010-03-03 12 +439 val_439 2010-03-03 12 +44 val_44 2010-03-03 12 +44 val_44 2010-03-03 12 +443 val_443 2010-03-03 12 +443 val_443 2010-03-03 12 +444 val_444 2010-03-03 12 +444 val_444 2010-03-03 12 +446 val_446 2010-03-03 12 +446 val_446 2010-03-03 12 +448 val_448 2010-03-03 12 +448 val_448 2010-03-03 12 +449 val_449 2010-03-03 12 +449 val_449 2010-03-03 12 +452 val_452 2010-03-03 12 +452 val_452 2010-03-03 12 +453 val_453 2010-03-03 12 +453 val_453 2010-03-03 12 +454 val_454 2010-03-03 12 +454 val_454 2010-03-03 12 +454 val_454 2010-03-03 12 +454 val_454 2010-03-03 12 +454 val_454 2010-03-03 12 +454 val_454 2010-03-03 12 +455 val_455 2010-03-03 12 +455 val_455 2010-03-03 12 +457 val_457 2010-03-03 12 +457 val_457 2010-03-03 12 +458 val_458 2010-03-03 12 +458 val_458 2010-03-03 12 +458 val_458 2010-03-03 12 +458 val_458 2010-03-03 12 +459 val_459 2010-03-03 12 +459 val_459 2010-03-03 12 +459 val_459 2010-03-03 12 +459 val_459 2010-03-03 12 +460 val_460 2010-03-03 12 +460 val_460 2010-03-03 12 +462 val_462 2010-03-03 12 +462 val_462 2010-03-03 12 +462 val_462 2010-03-03 12 +462 val_462 2010-03-03 12 +463 val_463 2010-03-03 12 +463 val_463 2010-03-03 12 +463 val_463 2010-03-03 12 463 val_463 2010-03-03 12 -24 val_24 2010-03-03 12 -348 val_348 2010-03-03 12 -208 val_208 2010-03-03 12 -230 val_230 2010-03-03 12 -411 val_411 2010-03-03 12 -8 val_8 2010-03-03 12 -58 val_58 2010-03-03 12 466 val_466 2010-03-03 12 -348 val_348 2010-03-03 12 -84 val_84 2010-03-03 12 -217 val_217 2010-03-03 12 -138 val_138 2010-03-03 12 -272 val_272 2010-03-03 12 -373 val_373 2010-03-03 12 -353 val_353 2010-03-03 12 -489 val_489 2010-03-03 12 -384 val_384 2010-03-03 12 -404 val_404 2010-03-03 12 -260 val_260 2010-03-03 12 -67 val_67 2010-03-03 12 -181 val_181 2010-03-03 12 -230 val_230 2010-03-03 12 -12 val_12 2010-03-03 12 -284 val_284 2010-03-03 12 -180 val_180 2010-03-03 12 -333 val_333 2010-03-03 12 -83 val_83 2010-03-03 12 -241 val_241 2010-03-03 12 -137 val_137 2010-03-03 12 -275 val_275 2010-03-03 12 -311 val_311 2010-03-03 12 -332 val_332 2010-03-03 12 -318 val_318 2010-03-03 12 -65 val_65 2010-03-03 12 -221 val_221 2010-03-03 12 -289 val_289 2010-03-03 12 -176 val_176 2010-03-03 12 -278 val_278 2010-03-03 12 -430 val_430 2010-03-03 12 -216 val_216 2010-03-03 12 -213 val_213 2010-03-03 12 -239 val_239 2010-03-03 12 -103 val_103 2010-03-03 12 -138 val_138 2010-03-03 12 -51 val_51 2010-03-03 12 -459 val_459 2010-03-03 12 -54 val_54 2010-03-03 12 -176 val_176 2010-03-03 12 -187 val_187 2010-03-03 12 -286 val_286 2010-03-03 12 -192 val_192 2010-03-03 12 +466 val_466 2010-03-03 12 +466 val_466 2010-03-03 12 +466 val_466 2010-03-03 12 +466 val_466 2010-03-03 12 +466 val_466 2010-03-03 12 +467 val_467 2010-03-03 12 +467 val_467 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +468 val_468 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 +469 val_469 2010-03-03 12 469 val_469 2010-03-03 12 -437 val_437 2010-03-03 12 -386 val_386 2010-03-03 12 -125 val_125 2010-03-03 12 -498 val_498 2010-03-03 12 -382 val_382 2010-03-03 12 -169 val_169 2010-03-03 12 -399 val_399 2010-03-03 12 -356 val_356 2010-03-03 12 -208 val_208 2010-03-03 12 -277 val_277 2010-03-03 12 -427 val_427 2010-03-03 12 -35 val_35 2010-03-03 12 -280 val_280 2010-03-03 12 -4 val_4 2010-03-03 12 -72 val_72 2010-03-03 12 47 val_47 2010-03-03 12 -111 val_111 2010-03-03 12 -92 val_92 2010-03-03 12 -221 val_221 2010-03-03 12 -378 val_378 2010-03-03 12 -157 val_157 2010-03-03 12 -489 val_489 2010-03-03 12 -20 val_20 2010-03-03 12 -170 val_170 2010-03-03 12 -129 val_129 2010-03-03 12 -345 val_345 2010-03-03 12 -438 val_438 2010-03-03 12 -149 val_149 2010-03-03 12 -205 val_205 2010-03-03 12 -302 val_302 2010-03-03 12 -57 val_57 2010-03-03 12 -316 val_316 2010-03-03 12 -311 val_311 2010-03-03 12 -128 val_128 2010-03-03 12 -455 val_455 2010-03-03 12 -0 val_0 2010-03-03 12 -339 val_339 2010-03-03 12 -203 val_203 2010-03-03 12 -155 val_155 2010-03-03 12 -113 val_113 2010-03-03 12 -17 val_17 2010-03-03 12 +47 val_47 2010-03-03 12 +470 val_470 2010-03-03 12 +470 val_470 2010-03-03 12 +472 val_472 2010-03-03 12 +472 val_472 2010-03-03 12 475 val_475 2010-03-03 12 -195 val_195 2010-03-03 12 -167 val_167 2010-03-03 12 -325 val_325 2010-03-03 12 -367 val_367 2010-03-03 12 -342 val_342 2010-03-03 12 -439 val_439 2010-03-03 12 -266 val_266 2010-03-03 12 -365 val_365 2010-03-03 12 -309 val_309 2010-03-03 12 -397 val_397 2010-03-03 12 -377 val_377 2010-03-03 12 -162 val_162 2010-03-03 12 +475 val_475 2010-03-03 12 +477 val_477 2010-03-03 12 +477 val_477 2010-03-03 12 +478 val_478 2010-03-03 12 +478 val_478 2010-03-03 12 +478 val_478 2010-03-03 12 +478 val_478 2010-03-03 12 +479 val_479 2010-03-03 12 +479 val_479 2010-03-03 12 +480 val_480 2010-03-03 12 +480 val_480 2010-03-03 12 +480 val_480 2010-03-03 12 +480 val_480 2010-03-03 12 +480 val_480 2010-03-03 12 +480 val_480 2010-03-03 12 +481 val_481 2010-03-03 12 +481 val_481 2010-03-03 12 +482 val_482 2010-03-03 12 +482 val_482 2010-03-03 12 +483 val_483 2010-03-03 12 +483 val_483 2010-03-03 12 +484 val_484 2010-03-03 12 +484 val_484 2010-03-03 12 +485 val_485 2010-03-03 12 +485 val_485 2010-03-03 12 +487 val_487 2010-03-03 12 +487 val_487 2010-03-03 12 489 val_489 2010-03-03 12 -417 val_417 2010-03-03 12 -247 val_247 2010-03-03 12 -396 val_396 2010-03-03 12 -399 val_399 2010-03-03 12 -174 val_174 2010-03-03 12 -208 val_208 2010-03-03 12 -466 val_466 2010-03-03 12 -199 val_199 2010-03-03 12 -207 val_207 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +489 val_489 2010-03-03 12 +490 val_490 2010-03-03 12 +490 val_490 2010-03-03 12 +491 val_491 2010-03-03 12 +491 val_491 2010-03-03 12 +492 val_492 2010-03-03 12 +492 val_492 2010-03-03 12 +492 val_492 2010-03-03 12 +492 val_492 2010-03-03 12 +493 val_493 2010-03-03 12 +493 val_493 2010-03-03 12 +494 val_494 2010-03-03 12 494 val_494 2010-03-03 12 -413 val_413 2010-03-03 12 -174 val_174 2010-03-03 12 -482 val_482 2010-03-03 12 -237 val_237 2010-03-03 12 -394 val_394 2010-03-03 12 -459 val_459 2010-03-03 12 -446 val_446 2010-03-03 12 -338 val_338 2010-03-03 12 -193 val_193 2010-03-03 12 -153 val_153 2010-03-03 12 -287 val_287 2010-03-03 12 -219 val_219 2010-03-03 12 -292 val_292 2010-03-03 12 -252 val_252 2010-03-03 12 -430 val_430 2010-03-03 12 -417 val_417 2010-03-03 12 -166 val_166 2010-03-03 12 -403 val_403 2010-03-03 12 -82 val_82 2010-03-03 12 -15 val_15 2010-03-03 12 -209 val_209 2010-03-03 12 -277 val_277 2010-03-03 12 -281 val_281 2010-03-03 12 -327 val_327 2010-03-03 12 -37 val_37 2010-03-03 12 495 val_495 2010-03-03 12 -145 val_145 2010-03-03 12 -469 val_469 2010-03-03 12 -152 val_152 2010-03-03 12 -374 val_374 2010-03-03 12 -429 val_429 2010-03-03 12 -406 val_406 2010-03-03 12 -146 val_146 2010-03-03 12 -213 val_213 2010-03-03 12 -128 val_128 2010-03-03 12 +495 val_495 2010-03-03 12 +496 val_496 2010-03-03 12 +496 val_496 2010-03-03 12 +497 val_497 2010-03-03 12 +497 val_497 2010-03-03 12 +498 val_498 2010-03-03 12 +498 val_498 2010-03-03 12 +498 val_498 2010-03-03 12 +498 val_498 2010-03-03 12 +498 val_498 2010-03-03 12 +498 val_498 2010-03-03 12 +5 val_5 2010-03-03 12 +5 val_5 2010-03-03 12 +5 val_5 2010-03-03 12 +5 val_5 2010-03-03 12 +5 val_5 2010-03-03 12 +5 val_5 2010-03-03 12 +51 val_51 2010-03-03 12 +51 val_51 2010-03-03 12 +51 val_51 2010-03-03 12 +51 val_51 2010-03-03 12 +53 val_53 2010-03-03 12 +53 val_53 2010-03-03 12 +54 val_54 2010-03-03 12 +54 val_54 2010-03-03 12 +57 val_57 2010-03-03 12 +57 val_57 2010-03-03 12 +58 val_58 2010-03-03 12 +58 val_58 2010-03-03 12 +58 val_58 2010-03-03 12 +58 val_58 2010-03-03 12 +64 val_64 2010-03-03 12 +64 val_64 2010-03-03 12 +65 val_65 2010-03-03 12 +65 val_65 2010-03-03 12 66 val_66 2010-03-03 12 -369 val_369 2010-03-03 12 -224 val_224 2010-03-03 12 -273 val_273 2010-03-03 12 -150 val_150 2010-03-03 12 -401 val_401 2010-03-03 12 -193 val_193 2010-03-03 12 -265 val_265 2010-03-03 12 -484 val_484 2010-03-03 12 -98 val_98 2010-03-03 12 -278 val_278 2010-03-03 12 -255 val_255 2010-03-03 12 -409 val_409 2010-03-03 12 -165 val_165 2010-03-03 12 -27 val_27 2010-03-03 12 -311 val_311 2010-03-03 12 +66 val_66 2010-03-03 12 +67 val_67 2010-03-03 12 +67 val_67 2010-03-03 12 +67 val_67 2010-03-03 12 +67 val_67 2010-03-03 12 +69 val_69 2010-03-03 12 +69 val_69 2010-03-03 12 +70 val_70 2010-03-03 12 +70 val_70 2010-03-03 12 +70 val_70 2010-03-03 12 +70 val_70 2010-03-03 12 +70 val_70 2010-03-03 12 +70 val_70 2010-03-03 12 +72 val_72 2010-03-03 12 +72 val_72 2010-03-03 12 +72 val_72 2010-03-03 12 +72 val_72 2010-03-03 12 +74 val_74 2010-03-03 12 +74 val_74 2010-03-03 12 +76 val_76 2010-03-03 12 +76 val_76 2010-03-03 12 +76 val_76 2010-03-03 12 +76 val_76 2010-03-03 12 +77 val_77 2010-03-03 12 +77 val_77 2010-03-03 12 +78 val_78 2010-03-03 12 +78 val_78 2010-03-03 12 +8 val_8 2010-03-03 12 +8 val_8 2010-03-03 12 +80 val_80 2010-03-03 12 +80 val_80 2010-03-03 12 +82 val_82 2010-03-03 12 +82 val_82 2010-03-03 12 +83 val_83 2010-03-03 12 +83 val_83 2010-03-03 12 +83 val_83 2010-03-03 12 +83 val_83 2010-03-03 12 +84 val_84 2010-03-03 12 +84 val_84 2010-03-03 12 +84 val_84 2010-03-03 12 +84 val_84 2010-03-03 12 +85 val_85 2010-03-03 12 +85 val_85 2010-03-03 12 86 val_86 2010-03-03 12 -238 val_238 2010-03-03 12 +86 val_86 2010-03-03 12 +87 val_87 2010-03-03 12 +87 val_87 2010-03-03 12 +9 val_9 2010-03-03 12 +9 val_9 2010-03-03 12 +90 val_90 2010-03-03 12 +90 val_90 2010-03-03 12 +90 val_90 2010-03-03 12 +90 val_90 2010-03-03 12 +90 val_90 2010-03-03 12 +90 val_90 2010-03-03 12 +92 val_92 2010-03-03 12 +92 val_92 2010-03-03 12 +95 val_95 2010-03-03 12 +95 val_95 2010-03-03 12 +95 val_95 2010-03-03 12 +95 val_95 2010-03-03 12 +96 val_96 2010-03-03 12 +96 val_96 2010-03-03 12 +97 val_97 2010-03-03 12 +97 val_97 2010-03-03 12 +97 val_97 2010-03-03 12 97 val_97 2010-03-03 12 +98 val_98 2010-03-03 12 +98 val_98 2010-03-03 12 +98 val_98 2010-03-03 12 +98 val_98 2010-03-03 12 diff --git ql/src/test/results/clientpositive/load_dyn_part12.q.out ql/src/test/results/clientpositive/llap/load_dyn_part12.q.out similarity index 100% rename from ql/src/test/results/clientpositive/load_dyn_part12.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part12.q.out index 471b44014b..01cdfd4cde 100644 --- ql/src/test/results/clientpositive/load_dyn_part12.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part12.q.out @@ -75,2002 +75,2002 @@ POSTHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=22 POSTHOOK: Input: default@nzhang_part12@ds=2010-03-03/hr=24 #### A masked pattern was here #### 238 val_238 2010-03-03 22 -97 val_97 2010-03-03 22 -200 val_200 2010-03-03 22 -400 val_400 2010-03-03 22 -403 val_403 2010-03-03 22 -169 val_169 2010-03-03 22 -90 val_90 2010-03-03 22 -126 val_126 2010-03-03 22 -222 val_222 2010-03-03 22 -477 val_477 2010-03-03 22 -414 val_414 2010-03-03 22 -194 val_194 2010-03-03 22 -307 val_307 2010-03-03 22 -348 val_348 2010-03-03 22 -152 val_152 2010-03-03 22 -448 val_448 2010-03-03 22 -37 val_37 2010-03-03 22 -28 val_28 2010-03-03 22 -84 val_84 2010-03-03 22 -315 val_315 2010-03-03 22 -469 val_469 2010-03-03 22 -97 val_97 2010-03-03 22 -344 val_344 2010-03-03 22 -281 val_281 2010-03-03 22 -183 val_183 2010-03-03 22 +86 val_86 2010-03-03 22 +311 val_311 2010-03-03 22 +27 val_27 2010-03-03 22 +165 val_165 2010-03-03 22 +409 val_409 2010-03-03 22 +255 val_255 2010-03-03 22 +278 val_278 2010-03-03 22 +98 val_98 2010-03-03 22 +484 val_484 2010-03-03 22 +265 val_265 2010-03-03 22 +193 val_193 2010-03-03 22 +401 val_401 2010-03-03 22 +150 val_150 2010-03-03 22 273 val_273 2010-03-03 22 -18 val_18 2010-03-03 22 -167 val_167 2010-03-03 22 -348 val_348 2010-03-03 22 -285 val_285 2010-03-03 22 -186 val_186 2010-03-03 22 -362 val_362 2010-03-03 22 -458 val_458 2010-03-03 22 +224 val_224 2010-03-03 22 +369 val_369 2010-03-03 22 +66 val_66 2010-03-03 22 +128 val_128 2010-03-03 22 +213 val_213 2010-03-03 22 146 val_146 2010-03-03 22 -498 val_498 2010-03-03 22 -341 val_341 2010-03-03 22 -9 val_9 2010-03-03 22 -298 val_298 2010-03-03 22 -100 val_100 2010-03-03 22 -492 val_492 2010-03-03 22 -462 val_462 2010-03-03 22 -18 val_18 2010-03-03 22 -379 val_379 2010-03-03 22 -384 val_384 2010-03-03 22 -67 val_67 2010-03-03 22 -134 val_134 2010-03-03 22 -26 val_26 2010-03-03 22 -256 val_256 2010-03-03 22 -384 val_384 2010-03-03 22 -407 val_407 2010-03-03 22 -421 val_421 2010-03-03 22 -401 val_401 2010-03-03 22 -375 val_375 2010-03-03 22 -454 val_454 2010-03-03 22 -189 val_189 2010-03-03 22 -175 val_175 2010-03-03 22 -133 val_133 2010-03-03 22 406 val_406 2010-03-03 22 -233 val_233 2010-03-03 22 -462 val_462 2010-03-03 22 -214 val_214 2010-03-03 22 -172 val_172 2010-03-03 22 -353 val_353 2010-03-03 22 -136 val_136 2010-03-03 22 -83 val_83 2010-03-03 22 -480 val_480 2010-03-03 22 -265 val_265 2010-03-03 22 -249 val_249 2010-03-03 22 -207 val_207 2010-03-03 22 -460 val_460 2010-03-03 22 -493 val_493 2010-03-03 22 -333 val_333 2010-03-03 22 -317 val_317 2010-03-03 22 -310 val_310 2010-03-03 22 -468 val_468 2010-03-03 22 -178 val_178 2010-03-03 22 -478 val_478 2010-03-03 22 -230 val_230 2010-03-03 22 -277 val_277 2010-03-03 22 -325 val_325 2010-03-03 22 -323 val_323 2010-03-03 22 -443 val_443 2010-03-03 22 -169 val_169 2010-03-03 22 429 val_429 2010-03-03 22 -120 val_120 2010-03-03 22 -444 val_444 2010-03-03 22 +374 val_374 2010-03-03 22 +152 val_152 2010-03-03 22 +469 val_469 2010-03-03 22 +145 val_145 2010-03-03 22 +495 val_495 2010-03-03 22 +37 val_37 2010-03-03 22 +327 val_327 2010-03-03 22 +281 val_281 2010-03-03 22 +277 val_277 2010-03-03 22 +209 val_209 2010-03-03 22 +15 val_15 2010-03-03 22 +82 val_82 2010-03-03 22 +403 val_403 2010-03-03 22 +166 val_166 2010-03-03 22 +417 val_417 2010-03-03 22 +430 val_430 2010-03-03 22 +252 val_252 2010-03-03 22 +292 val_292 2010-03-03 22 +219 val_219 2010-03-03 22 +287 val_287 2010-03-03 22 +153 val_153 2010-03-03 22 +193 val_193 2010-03-03 22 +338 val_338 2010-03-03 22 +446 val_446 2010-03-03 22 +459 val_459 2010-03-03 22 +394 val_394 2010-03-03 22 +237 val_237 2010-03-03 22 +482 val_482 2010-03-03 22 +174 val_174 2010-03-03 22 +413 val_413 2010-03-03 22 +494 val_494 2010-03-03 22 +207 val_207 2010-03-03 22 199 val_199 2010-03-03 22 +466 val_466 2010-03-03 22 +208 val_208 2010-03-03 22 +174 val_174 2010-03-03 22 +399 val_399 2010-03-03 22 +396 val_396 2010-03-03 22 +247 val_247 2010-03-03 22 417 val_417 2010-03-03 22 -305 val_305 2010-03-03 22 -479 val_479 2010-03-03 22 -248 val_248 2010-03-03 22 -360 val_360 2010-03-03 22 +489 val_489 2010-03-03 22 +162 val_162 2010-03-03 22 +377 val_377 2010-03-03 22 +397 val_397 2010-03-03 22 +309 val_309 2010-03-03 22 +365 val_365 2010-03-03 22 +266 val_266 2010-03-03 22 439 val_439 2010-03-03 22 -237 val_237 2010-03-03 22 -491 val_491 2010-03-03 22 -200 val_200 2010-03-03 22 -414 val_414 2010-03-03 22 -119 val_119 2010-03-03 22 +342 val_342 2010-03-03 22 +367 val_367 2010-03-03 22 +325 val_325 2010-03-03 22 +167 val_167 2010-03-03 22 +195 val_195 2010-03-03 22 +475 val_475 2010-03-03 22 +17 val_17 2010-03-03 22 +113 val_113 2010-03-03 22 +155 val_155 2010-03-03 22 +203 val_203 2010-03-03 22 +339 val_339 2010-03-03 22 +0 val_0 2010-03-03 22 +455 val_455 2010-03-03 22 +128 val_128 2010-03-03 22 +311 val_311 2010-03-03 22 +316 val_316 2010-03-03 22 +57 val_57 2010-03-03 22 +302 val_302 2010-03-03 22 +205 val_205 2010-03-03 22 +149 val_149 2010-03-03 22 438 val_438 2010-03-03 22 -163 val_163 2010-03-03 22 -70 val_70 2010-03-03 22 -104 val_104 2010-03-03 22 -255 val_255 2010-03-03 22 -351 val_351 2010-03-03 22 -24 val_24 2010-03-03 22 -291 val_291 2010-03-03 22 -480 val_480 2010-03-03 22 -397 val_397 2010-03-03 22 -70 val_70 2010-03-03 22 -5 val_5 2010-03-03 22 +345 val_345 2010-03-03 22 +129 val_129 2010-03-03 22 +170 val_170 2010-03-03 22 +20 val_20 2010-03-03 22 +489 val_489 2010-03-03 22 +157 val_157 2010-03-03 22 +378 val_378 2010-03-03 22 +221 val_221 2010-03-03 22 +92 val_92 2010-03-03 22 +111 val_111 2010-03-03 22 +47 val_47 2010-03-03 22 +72 val_72 2010-03-03 22 +4 val_4 2010-03-03 22 +280 val_280 2010-03-03 22 +35 val_35 2010-03-03 22 +427 val_427 2010-03-03 22 +277 val_277 2010-03-03 22 +208 val_208 2010-03-03 22 +356 val_356 2010-03-03 22 +399 val_399 2010-03-03 22 +169 val_169 2010-03-03 22 382 val_382 2010-03-03 22 -187 val_187 2010-03-03 22 -424 val_424 2010-03-03 22 -164 val_164 2010-03-03 22 -431 val_431 2010-03-03 22 +498 val_498 2010-03-03 22 125 val_125 2010-03-03 22 -298 val_298 2010-03-03 22 -478 val_478 2010-03-03 22 -454 val_454 2010-03-03 22 -431 val_431 2010-03-03 22 -164 val_164 2010-03-03 22 -217 val_217 2010-03-03 22 -201 val_201 2010-03-03 22 -396 val_396 2010-03-03 22 -12 val_12 2010-03-03 22 -424 val_424 2010-03-03 22 -348 val_348 2010-03-03 22 -262 val_262 2010-03-03 22 -203 val_203 2010-03-03 22 -90 val_90 2010-03-03 22 -258 val_258 2010-03-03 22 -114 val_114 2010-03-03 22 -401 val_401 2010-03-03 22 -406 val_406 2010-03-03 22 -190 val_190 2010-03-03 22 -409 val_409 2010-03-03 22 -406 val_406 2010-03-03 22 -257 val_257 2010-03-03 22 -105 val_105 2010-03-03 22 -53 val_53 2010-03-03 22 -483 val_483 2010-03-03 22 -403 val_403 2010-03-03 22 -175 val_175 2010-03-03 22 -366 val_366 2010-03-03 22 -466 val_466 2010-03-03 22 -104 val_104 2010-03-03 22 -335 val_335 2010-03-03 22 -321 val_321 2010-03-03 22 -193 val_193 2010-03-03 22 -44 val_44 2010-03-03 22 -80 val_80 2010-03-03 22 -235 val_235 2010-03-03 22 -331 val_331 2010-03-03 22 -283 val_283 2010-03-03 22 -35 val_35 2010-03-03 22 -2 val_2 2010-03-03 22 -280 val_280 2010-03-03 22 -463 val_463 2010-03-03 22 +386 val_386 2010-03-03 22 +437 val_437 2010-03-03 22 469 val_469 2010-03-03 22 -229 val_229 2010-03-03 22 -316 val_316 2010-03-03 22 -202 val_202 2010-03-03 22 -432 val_432 2010-03-03 22 -467 val_467 2010-03-03 22 -128 val_128 2010-03-03 22 -438 val_438 2010-03-03 22 -244 val_244 2010-03-03 22 -5 val_5 2010-03-03 22 -191 val_191 2010-03-03 22 -288 val_288 2010-03-03 22 -401 val_401 2010-03-03 22 -480 val_480 2010-03-03 22 -487 val_487 2010-03-03 22 -70 val_70 2010-03-03 22 -263 val_263 2010-03-03 22 -256 val_256 2010-03-03 22 -223 val_223 2010-03-03 22 -116 val_116 2010-03-03 22 -485 val_485 2010-03-03 22 +192 val_192 2010-03-03 22 +286 val_286 2010-03-03 22 +187 val_187 2010-03-03 22 +176 val_176 2010-03-03 22 +54 val_54 2010-03-03 22 +459 val_459 2010-03-03 22 +51 val_51 2010-03-03 22 +138 val_138 2010-03-03 22 +103 val_103 2010-03-03 22 239 val_239 2010-03-03 22 -219 val_219 2010-03-03 22 -274 val_274 2010-03-03 22 -167 val_167 2010-03-03 22 -344 val_344 2010-03-03 22 -367 val_367 2010-03-03 22 +213 val_213 2010-03-03 22 216 val_216 2010-03-03 22 -113 val_113 2010-03-03 22 -296 val_296 2010-03-03 22 -103 val_103 2010-03-03 22 -368 val_368 2010-03-03 22 -33 val_33 2010-03-03 22 +430 val_430 2010-03-03 22 +278 val_278 2010-03-03 22 +176 val_176 2010-03-03 22 +289 val_289 2010-03-03 22 +221 val_221 2010-03-03 22 +65 val_65 2010-03-03 22 +318 val_318 2010-03-03 22 +332 val_332 2010-03-03 22 +311 val_311 2010-03-03 22 +275 val_275 2010-03-03 22 +137 val_137 2010-03-03 22 +241 val_241 2010-03-03 22 +83 val_83 2010-03-03 22 +333 val_333 2010-03-03 22 +180 val_180 2010-03-03 22 +284 val_284 2010-03-03 22 +12 val_12 2010-03-03 22 230 val_230 2010-03-03 22 -69 val_69 2010-03-03 22 -342 val_342 2010-03-03 22 -74 val_74 2010-03-03 22 -76 val_76 2010-03-03 22 -468 val_468 2010-03-03 22 -64 val_64 2010-03-03 22 -209 val_209 2010-03-03 22 -30 val_30 2010-03-03 22 -453 val_453 2010-03-03 22 -138 val_138 2010-03-03 22 -228 val_228 2010-03-03 22 -218 val_218 2010-03-03 22 -449 val_449 2010-03-03 22 -149 val_149 2010-03-03 22 -492 val_492 2010-03-03 22 -223 val_223 2010-03-03 22 -41 val_41 2010-03-03 22 -76 val_76 2010-03-03 22 -78 val_78 2010-03-03 22 -458 val_458 2010-03-03 22 +181 val_181 2010-03-03 22 +67 val_67 2010-03-03 22 +260 val_260 2010-03-03 22 +404 val_404 2010-03-03 22 +384 val_384 2010-03-03 22 489 val_489 2010-03-03 22 -119 val_119 2010-03-03 22 -430 val_430 2010-03-03 22 -321 val_321 2010-03-03 22 -42 val_42 2010-03-03 22 -195 val_195 2010-03-03 22 -160 val_160 2010-03-03 22 -498 val_498 2010-03-03 22 -322 val_322 2010-03-03 22 -472 val_472 2010-03-03 22 -143 val_143 2010-03-03 22 -233 val_233 2010-03-03 22 -229 val_229 2010-03-03 22 -34 val_34 2010-03-03 22 -168 val_168 2010-03-03 22 -11 val_11 2010-03-03 22 -95 val_95 2010-03-03 22 -336 val_336 2010-03-03 22 -35 val_35 2010-03-03 22 -58 val_58 2010-03-03 22 -395 val_395 2010-03-03 22 -317 val_317 2010-03-03 22 -396 val_396 2010-03-03 22 -402 val_402 2010-03-03 22 -497 val_497 2010-03-03 22 -5 val_5 2010-03-03 22 -226 val_226 2010-03-03 22 -177 val_177 2010-03-03 22 -452 val_452 2010-03-03 22 -242 val_242 2010-03-03 22 -401 val_401 2010-03-03 22 -331 val_331 2010-03-03 22 +353 val_353 2010-03-03 22 +373 val_373 2010-03-03 22 272 val_272 2010-03-03 22 -392 val_392 2010-03-03 22 -369 val_369 2010-03-03 22 -242 val_242 2010-03-03 22 -327 val_327 2010-03-03 22 -389 val_389 2010-03-03 22 -309 val_309 2010-03-03 22 -224 val_224 2010-03-03 22 -306 val_306 2010-03-03 22 -273 val_273 2010-03-03 22 -277 val_277 2010-03-03 22 -10 val_10 2010-03-03 22 -435 val_435 2010-03-03 22 -19 val_19 2010-03-03 22 -307 val_307 2010-03-03 22 -90 val_90 2010-03-03 22 -72 val_72 2010-03-03 22 -118 val_118 2010-03-03 22 -15 val_15 2010-03-03 22 -419 val_419 2010-03-03 22 -238 val_238 2010-03-03 22 138 val_138 2010-03-03 22 -282 val_282 2010-03-03 22 -395 val_395 2010-03-03 22 -134 val_134 2010-03-03 22 -118 val_118 2010-03-03 22 +217 val_217 2010-03-03 22 +84 val_84 2010-03-03 22 +348 val_348 2010-03-03 22 +466 val_466 2010-03-03 22 +58 val_58 2010-03-03 22 +8 val_8 2010-03-03 22 +411 val_411 2010-03-03 22 +230 val_230 2010-03-03 22 +208 val_208 2010-03-03 22 +348 val_348 2010-03-03 22 +24 val_24 2010-03-03 22 +463 val_463 2010-03-03 22 +431 val_431 2010-03-03 22 179 val_179 2010-03-03 22 -364 val_364 2010-03-03 22 -87 val_87 2010-03-03 22 -490 val_490 2010-03-03 22 +172 val_172 2010-03-03 22 +42 val_42 2010-03-03 22 +129 val_129 2010-03-03 22 +158 val_158 2010-03-03 22 +119 val_119 2010-03-03 22 +496 val_496 2010-03-03 22 0 val_0 2010-03-03 22 -77 val_77 2010-03-03 22 -85 val_85 2010-03-03 22 -413 val_413 2010-03-03 22 -169 val_169 2010-03-03 22 -316 val_316 2010-03-03 22 -369 val_369 2010-03-03 22 -137 val_137 2010-03-03 22 -470 val_470 2010-03-03 22 -409 val_409 2010-03-03 22 -318 val_318 2010-03-03 22 -318 val_318 2010-03-03 22 -187 val_187 2010-03-03 22 +322 val_322 2010-03-03 22 197 val_197 2010-03-03 22 -282 val_282 2010-03-03 22 -98 val_98 2010-03-03 22 -457 val_457 2010-03-03 22 -481 val_481 2010-03-03 22 -288 val_288 2010-03-03 22 -196 val_196 2010-03-03 22 -95 val_95 2010-03-03 22 -308 val_308 2010-03-03 22 468 val_468 2010-03-03 22 -469 val_469 2010-03-03 22 -156 val_156 2010-03-03 22 -436 val_436 2010-03-03 22 -43 val_43 2010-03-03 22 -404 val_404 2010-03-03 22 -51 val_51 2010-03-03 22 -131 val_131 2010-03-03 22 -120 val_120 2010-03-03 22 -205 val_205 2010-03-03 22 -230 val_230 2010-03-03 22 -327 val_327 2010-03-03 22 -165 val_165 2010-03-03 22 -26 val_26 2010-03-03 22 -96 val_96 2010-03-03 22 -418 val_418 2010-03-03 22 -191 val_191 2010-03-03 22 -199 val_199 2010-03-03 22 -298 val_298 2010-03-03 22 -100 val_100 2010-03-03 22 -454 val_454 2010-03-03 22 393 val_393 2010-03-03 22 +454 val_454 2010-03-03 22 +100 val_100 2010-03-03 22 +298 val_298 2010-03-03 22 +199 val_199 2010-03-03 22 +191 val_191 2010-03-03 22 +418 val_418 2010-03-03 22 +96 val_96 2010-03-03 22 +26 val_26 2010-03-03 22 +165 val_165 2010-03-03 22 +327 val_327 2010-03-03 22 +230 val_230 2010-03-03 22 +205 val_205 2010-03-03 22 +120 val_120 2010-03-03 22 +131 val_131 2010-03-03 22 +51 val_51 2010-03-03 22 +404 val_404 2010-03-03 22 +43 val_43 2010-03-03 22 +436 val_436 2010-03-03 22 +156 val_156 2010-03-03 22 +469 val_469 2010-03-03 22 468 val_468 2010-03-03 22 +308 val_308 2010-03-03 22 +95 val_95 2010-03-03 22 +196 val_196 2010-03-03 22 +288 val_288 2010-03-03 22 +481 val_481 2010-03-03 22 +457 val_457 2010-03-03 22 +98 val_98 2010-03-03 22 +282 val_282 2010-03-03 22 197 val_197 2010-03-03 22 -322 val_322 2010-03-03 22 +187 val_187 2010-03-03 22 +318 val_318 2010-03-03 22 +318 val_318 2010-03-03 22 +409 val_409 2010-03-03 22 +470 val_470 2010-03-03 22 +137 val_137 2010-03-03 22 +369 val_369 2010-03-03 22 +316 val_316 2010-03-03 22 +169 val_169 2010-03-03 22 +413 val_413 2010-03-03 22 +85 val_85 2010-03-03 22 +77 val_77 2010-03-03 22 0 val_0 2010-03-03 22 -496 val_496 2010-03-03 22 -119 val_119 2010-03-03 22 -158 val_158 2010-03-03 22 -129 val_129 2010-03-03 22 -42 val_42 2010-03-03 22 -172 val_172 2010-03-03 22 +490 val_490 2010-03-03 22 +87 val_87 2010-03-03 22 +364 val_364 2010-03-03 22 179 val_179 2010-03-03 22 -431 val_431 2010-03-03 22 -463 val_463 2010-03-03 22 -24 val_24 2010-03-03 22 -348 val_348 2010-03-03 22 -208 val_208 2010-03-03 22 -230 val_230 2010-03-03 22 -411 val_411 2010-03-03 22 -8 val_8 2010-03-03 22 -58 val_58 2010-03-03 22 -466 val_466 2010-03-03 22 -348 val_348 2010-03-03 22 -84 val_84 2010-03-03 22 -217 val_217 2010-03-03 22 +118 val_118 2010-03-03 22 +134 val_134 2010-03-03 22 +395 val_395 2010-03-03 22 +282 val_282 2010-03-03 22 138 val_138 2010-03-03 22 +238 val_238 2010-03-03 22 +419 val_419 2010-03-03 22 +15 val_15 2010-03-03 22 +118 val_118 2010-03-03 22 +72 val_72 2010-03-03 22 +90 val_90 2010-03-03 22 +307 val_307 2010-03-03 22 +19 val_19 2010-03-03 22 +435 val_435 2010-03-03 22 +10 val_10 2010-03-03 22 +277 val_277 2010-03-03 22 +273 val_273 2010-03-03 22 +306 val_306 2010-03-03 22 +224 val_224 2010-03-03 22 +309 val_309 2010-03-03 22 +389 val_389 2010-03-03 22 +327 val_327 2010-03-03 22 +242 val_242 2010-03-03 22 +369 val_369 2010-03-03 22 +392 val_392 2010-03-03 22 272 val_272 2010-03-03 22 -373 val_373 2010-03-03 22 -353 val_353 2010-03-03 22 +331 val_331 2010-03-03 22 +401 val_401 2010-03-03 22 +242 val_242 2010-03-03 22 +452 val_452 2010-03-03 22 +177 val_177 2010-03-03 22 +226 val_226 2010-03-03 22 +5 val_5 2010-03-03 22 +497 val_497 2010-03-03 22 +402 val_402 2010-03-03 22 +396 val_396 2010-03-03 22 +317 val_317 2010-03-03 22 +395 val_395 2010-03-03 22 +58 val_58 2010-03-03 22 +35 val_35 2010-03-03 22 +336 val_336 2010-03-03 22 +95 val_95 2010-03-03 22 +11 val_11 2010-03-03 22 +168 val_168 2010-03-03 22 +34 val_34 2010-03-03 22 +229 val_229 2010-03-03 22 +233 val_233 2010-03-03 22 +143 val_143 2010-03-03 22 +472 val_472 2010-03-03 22 +322 val_322 2010-03-03 22 +498 val_498 2010-03-03 22 +160 val_160 2010-03-03 22 +195 val_195 2010-03-03 22 +42 val_42 2010-03-03 22 +321 val_321 2010-03-03 22 +430 val_430 2010-03-03 22 +119 val_119 2010-03-03 22 489 val_489 2010-03-03 22 -384 val_384 2010-03-03 22 -404 val_404 2010-03-03 22 -260 val_260 2010-03-03 22 -67 val_67 2010-03-03 22 -181 val_181 2010-03-03 22 +458 val_458 2010-03-03 22 +78 val_78 2010-03-03 22 +76 val_76 2010-03-03 22 +41 val_41 2010-03-03 22 +223 val_223 2010-03-03 22 +492 val_492 2010-03-03 22 +149 val_149 2010-03-03 22 +449 val_449 2010-03-03 22 +218 val_218 2010-03-03 22 +228 val_228 2010-03-03 22 +138 val_138 2010-03-03 22 +453 val_453 2010-03-03 22 +30 val_30 2010-03-03 22 +209 val_209 2010-03-03 22 +64 val_64 2010-03-03 22 +468 val_468 2010-03-03 22 +76 val_76 2010-03-03 22 +74 val_74 2010-03-03 22 +342 val_342 2010-03-03 22 +69 val_69 2010-03-03 22 230 val_230 2010-03-03 22 -12 val_12 2010-03-03 22 -284 val_284 2010-03-03 22 -180 val_180 2010-03-03 22 -333 val_333 2010-03-03 22 -83 val_83 2010-03-03 22 -241 val_241 2010-03-03 22 -137 val_137 2010-03-03 22 -275 val_275 2010-03-03 22 -311 val_311 2010-03-03 22 -332 val_332 2010-03-03 22 -318 val_318 2010-03-03 22 -65 val_65 2010-03-03 22 -221 val_221 2010-03-03 22 -289 val_289 2010-03-03 22 -176 val_176 2010-03-03 22 -278 val_278 2010-03-03 22 -430 val_430 2010-03-03 22 +33 val_33 2010-03-03 22 +368 val_368 2010-03-03 22 +103 val_103 2010-03-03 22 +296 val_296 2010-03-03 22 +113 val_113 2010-03-03 22 216 val_216 2010-03-03 22 -213 val_213 2010-03-03 22 +367 val_367 2010-03-03 22 +344 val_344 2010-03-03 22 +167 val_167 2010-03-03 22 +274 val_274 2010-03-03 22 +219 val_219 2010-03-03 22 239 val_239 2010-03-03 22 -103 val_103 2010-03-03 22 -138 val_138 2010-03-03 22 -51 val_51 2010-03-03 22 -459 val_459 2010-03-03 22 -54 val_54 2010-03-03 22 -176 val_176 2010-03-03 22 -187 val_187 2010-03-03 22 -286 val_286 2010-03-03 22 -192 val_192 2010-03-03 22 +485 val_485 2010-03-03 22 +116 val_116 2010-03-03 22 +223 val_223 2010-03-03 22 +256 val_256 2010-03-03 22 +263 val_263 2010-03-03 22 +70 val_70 2010-03-03 22 +487 val_487 2010-03-03 22 +480 val_480 2010-03-03 22 +401 val_401 2010-03-03 22 +288 val_288 2010-03-03 22 +191 val_191 2010-03-03 22 +5 val_5 2010-03-03 22 +244 val_244 2010-03-03 22 +438 val_438 2010-03-03 22 +128 val_128 2010-03-03 22 +467 val_467 2010-03-03 22 +432 val_432 2010-03-03 22 +202 val_202 2010-03-03 22 +316 val_316 2010-03-03 22 +229 val_229 2010-03-03 22 469 val_469 2010-03-03 22 -437 val_437 2010-03-03 22 -386 val_386 2010-03-03 22 +463 val_463 2010-03-03 22 +280 val_280 2010-03-03 22 +2 val_2 2010-03-03 22 +35 val_35 2010-03-03 22 +283 val_283 2010-03-03 22 +331 val_331 2010-03-03 22 +235 val_235 2010-03-03 22 +80 val_80 2010-03-03 22 +44 val_44 2010-03-03 22 +193 val_193 2010-03-03 22 +321 val_321 2010-03-03 22 +335 val_335 2010-03-03 22 +104 val_104 2010-03-03 22 +466 val_466 2010-03-03 22 +366 val_366 2010-03-03 22 +175 val_175 2010-03-03 22 +403 val_403 2010-03-03 22 +483 val_483 2010-03-03 22 +53 val_53 2010-03-03 22 +105 val_105 2010-03-03 22 +257 val_257 2010-03-03 22 +406 val_406 2010-03-03 22 +409 val_409 2010-03-03 22 +190 val_190 2010-03-03 22 +406 val_406 2010-03-03 22 +401 val_401 2010-03-03 22 +114 val_114 2010-03-03 22 +258 val_258 2010-03-03 22 +90 val_90 2010-03-03 22 +203 val_203 2010-03-03 22 +262 val_262 2010-03-03 22 +348 val_348 2010-03-03 22 +424 val_424 2010-03-03 22 +12 val_12 2010-03-03 22 +396 val_396 2010-03-03 22 +201 val_201 2010-03-03 22 +217 val_217 2010-03-03 22 +164 val_164 2010-03-03 22 +431 val_431 2010-03-03 22 +454 val_454 2010-03-03 22 +478 val_478 2010-03-03 22 +298 val_298 2010-03-03 22 125 val_125 2010-03-03 22 -498 val_498 2010-03-03 22 +431 val_431 2010-03-03 22 +164 val_164 2010-03-03 22 +424 val_424 2010-03-03 22 +187 val_187 2010-03-03 22 382 val_382 2010-03-03 22 -169 val_169 2010-03-03 22 -399 val_399 2010-03-03 22 -356 val_356 2010-03-03 22 -208 val_208 2010-03-03 22 -277 val_277 2010-03-03 22 -427 val_427 2010-03-03 22 -35 val_35 2010-03-03 22 -280 val_280 2010-03-03 22 -4 val_4 2010-03-03 22 -72 val_72 2010-03-03 22 -47 val_47 2010-03-03 22 -111 val_111 2010-03-03 22 -92 val_92 2010-03-03 22 -221 val_221 2010-03-03 22 -378 val_378 2010-03-03 22 -157 val_157 2010-03-03 22 -489 val_489 2010-03-03 22 -20 val_20 2010-03-03 22 -170 val_170 2010-03-03 22 -129 val_129 2010-03-03 22 -345 val_345 2010-03-03 22 +5 val_5 2010-03-03 22 +70 val_70 2010-03-03 22 +397 val_397 2010-03-03 22 +480 val_480 2010-03-03 22 +291 val_291 2010-03-03 22 +24 val_24 2010-03-03 22 +351 val_351 2010-03-03 22 +255 val_255 2010-03-03 22 +104 val_104 2010-03-03 22 +70 val_70 2010-03-03 22 +163 val_163 2010-03-03 22 438 val_438 2010-03-03 22 -149 val_149 2010-03-03 22 -205 val_205 2010-03-03 22 -302 val_302 2010-03-03 22 -57 val_57 2010-03-03 22 -316 val_316 2010-03-03 22 -311 val_311 2010-03-03 22 -128 val_128 2010-03-03 22 -455 val_455 2010-03-03 22 -0 val_0 2010-03-03 22 -339 val_339 2010-03-03 22 -203 val_203 2010-03-03 22 -155 val_155 2010-03-03 22 -113 val_113 2010-03-03 22 -17 val_17 2010-03-03 22 -475 val_475 2010-03-03 22 -195 val_195 2010-03-03 22 -167 val_167 2010-03-03 22 -325 val_325 2010-03-03 22 -367 val_367 2010-03-03 22 -342 val_342 2010-03-03 22 +119 val_119 2010-03-03 22 +414 val_414 2010-03-03 22 +200 val_200 2010-03-03 22 +491 val_491 2010-03-03 22 +237 val_237 2010-03-03 22 439 val_439 2010-03-03 22 -266 val_266 2010-03-03 22 -365 val_365 2010-03-03 22 -309 val_309 2010-03-03 22 -397 val_397 2010-03-03 22 -377 val_377 2010-03-03 22 -162 val_162 2010-03-03 22 -489 val_489 2010-03-03 22 +360 val_360 2010-03-03 22 +248 val_248 2010-03-03 22 +479 val_479 2010-03-03 22 +305 val_305 2010-03-03 22 417 val_417 2010-03-03 22 -247 val_247 2010-03-03 22 -396 val_396 2010-03-03 22 -399 val_399 2010-03-03 22 -174 val_174 2010-03-03 22 -208 val_208 2010-03-03 22 -466 val_466 2010-03-03 22 199 val_199 2010-03-03 22 -207 val_207 2010-03-03 22 -494 val_494 2010-03-03 22 -413 val_413 2010-03-03 22 -174 val_174 2010-03-03 22 -482 val_482 2010-03-03 22 -237 val_237 2010-03-03 22 -394 val_394 2010-03-03 22 -459 val_459 2010-03-03 22 -446 val_446 2010-03-03 22 -338 val_338 2010-03-03 22 -193 val_193 2010-03-03 22 -153 val_153 2010-03-03 22 -287 val_287 2010-03-03 22 -219 val_219 2010-03-03 22 -292 val_292 2010-03-03 22 -252 val_252 2010-03-03 22 -430 val_430 2010-03-03 22 -417 val_417 2010-03-03 22 -166 val_166 2010-03-03 22 -403 val_403 2010-03-03 22 -82 val_82 2010-03-03 22 -15 val_15 2010-03-03 22 -209 val_209 2010-03-03 22 -277 val_277 2010-03-03 22 -281 val_281 2010-03-03 22 -327 val_327 2010-03-03 22 -37 val_37 2010-03-03 22 -495 val_495 2010-03-03 22 -145 val_145 2010-03-03 22 -469 val_469 2010-03-03 22 -152 val_152 2010-03-03 22 -374 val_374 2010-03-03 22 +444 val_444 2010-03-03 22 +120 val_120 2010-03-03 22 429 val_429 2010-03-03 22 +169 val_169 2010-03-03 22 +443 val_443 2010-03-03 22 +323 val_323 2010-03-03 22 +325 val_325 2010-03-03 22 +277 val_277 2010-03-03 22 +230 val_230 2010-03-03 22 +478 val_478 2010-03-03 22 +178 val_178 2010-03-03 22 +468 val_468 2010-03-03 22 +310 val_310 2010-03-03 22 +317 val_317 2010-03-03 22 +333 val_333 2010-03-03 22 +493 val_493 2010-03-03 22 +460 val_460 2010-03-03 22 +207 val_207 2010-03-03 22 +249 val_249 2010-03-03 22 +265 val_265 2010-03-03 22 +480 val_480 2010-03-03 22 +83 val_83 2010-03-03 22 +136 val_136 2010-03-03 22 +353 val_353 2010-03-03 22 +172 val_172 2010-03-03 22 +214 val_214 2010-03-03 22 +462 val_462 2010-03-03 22 +233 val_233 2010-03-03 22 406 val_406 2010-03-03 22 +133 val_133 2010-03-03 22 +175 val_175 2010-03-03 22 +189 val_189 2010-03-03 22 +454 val_454 2010-03-03 22 +375 val_375 2010-03-03 22 +401 val_401 2010-03-03 22 +421 val_421 2010-03-03 22 +407 val_407 2010-03-03 22 +384 val_384 2010-03-03 22 +256 val_256 2010-03-03 22 +26 val_26 2010-03-03 22 +134 val_134 2010-03-03 22 +67 val_67 2010-03-03 22 +384 val_384 2010-03-03 22 +379 val_379 2010-03-03 22 +18 val_18 2010-03-03 22 +462 val_462 2010-03-03 22 +492 val_492 2010-03-03 22 +100 val_100 2010-03-03 22 +298 val_298 2010-03-03 22 +9 val_9 2010-03-03 22 +341 val_341 2010-03-03 22 +498 val_498 2010-03-03 22 146 val_146 2010-03-03 22 -213 val_213 2010-03-03 22 -128 val_128 2010-03-03 22 -66 val_66 2010-03-03 22 -369 val_369 2010-03-03 22 -224 val_224 2010-03-03 22 +458 val_458 2010-03-03 22 +362 val_362 2010-03-03 22 +186 val_186 2010-03-03 22 +285 val_285 2010-03-03 22 +348 val_348 2010-03-03 22 +167 val_167 2010-03-03 22 +18 val_18 2010-03-03 22 273 val_273 2010-03-03 22 -150 val_150 2010-03-03 22 -401 val_401 2010-03-03 22 -193 val_193 2010-03-03 22 -265 val_265 2010-03-03 22 -484 val_484 2010-03-03 22 -98 val_98 2010-03-03 22 -278 val_278 2010-03-03 22 -255 val_255 2010-03-03 22 -409 val_409 2010-03-03 22 -165 val_165 2010-03-03 22 -27 val_27 2010-03-03 22 -311 val_311 2010-03-03 22 -86 val_86 2010-03-03 22 +183 val_183 2010-03-03 22 +281 val_281 2010-03-03 22 +344 val_344 2010-03-03 22 97 val_97 2010-03-03 22 -200 val_200 2010-03-03 22 -400 val_400 2010-03-03 22 -403 val_403 2010-03-03 22 -169 val_169 2010-03-03 22 -90 val_90 2010-03-03 22 -126 val_126 2010-03-03 22 -222 val_222 2010-03-03 22 -477 val_477 2010-03-03 22 -414 val_414 2010-03-03 22 -194 val_194 2010-03-03 22 -307 val_307 2010-03-03 22 -348 val_348 2010-03-03 22 -152 val_152 2010-03-03 22 -448 val_448 2010-03-03 22 -37 val_37 2010-03-03 22 -28 val_28 2010-03-03 22 -84 val_84 2010-03-03 22 -315 val_315 2010-03-03 22 469 val_469 2010-03-03 22 +315 val_315 2010-03-03 22 +84 val_84 2010-03-03 22 +28 val_28 2010-03-03 22 +37 val_37 2010-03-03 22 +448 val_448 2010-03-03 22 +152 val_152 2010-03-03 22 +348 val_348 2010-03-03 22 +307 val_307 2010-03-03 22 +194 val_194 2010-03-03 22 +414 val_414 2010-03-03 22 +477 val_477 2010-03-03 22 +222 val_222 2010-03-03 22 +126 val_126 2010-03-03 22 +90 val_90 2010-03-03 22 +169 val_169 2010-03-03 22 +403 val_403 2010-03-03 22 +400 val_400 2010-03-03 22 +200 val_200 2010-03-03 22 97 val_97 2010-03-03 22 -344 val_344 2010-03-03 22 -281 val_281 2010-03-03 22 -183 val_183 2010-03-03 22 +238 val_238 2010-03-03 22 +86 val_86 2010-03-03 22 +311 val_311 2010-03-03 22 +27 val_27 2010-03-03 22 +165 val_165 2010-03-03 22 +409 val_409 2010-03-03 22 +255 val_255 2010-03-03 22 +278 val_278 2010-03-03 22 +98 val_98 2010-03-03 22 +484 val_484 2010-03-03 22 +265 val_265 2010-03-03 22 +193 val_193 2010-03-03 22 +401 val_401 2010-03-03 22 +150 val_150 2010-03-03 22 273 val_273 2010-03-03 22 -18 val_18 2010-03-03 22 -167 val_167 2010-03-03 22 -348 val_348 2010-03-03 22 -285 val_285 2010-03-03 22 -186 val_186 2010-03-03 22 -362 val_362 2010-03-03 22 -458 val_458 2010-03-03 22 +224 val_224 2010-03-03 22 +369 val_369 2010-03-03 22 +66 val_66 2010-03-03 22 +128 val_128 2010-03-03 22 +213 val_213 2010-03-03 22 146 val_146 2010-03-03 22 -498 val_498 2010-03-03 22 -341 val_341 2010-03-03 22 -9 val_9 2010-03-03 22 -298 val_298 2010-03-03 22 -100 val_100 2010-03-03 22 -492 val_492 2010-03-03 22 -462 val_462 2010-03-03 22 -18 val_18 2010-03-03 22 -379 val_379 2010-03-03 22 -384 val_384 2010-03-03 22 -67 val_67 2010-03-03 22 -134 val_134 2010-03-03 22 -26 val_26 2010-03-03 22 -256 val_256 2010-03-03 22 -384 val_384 2010-03-03 22 -407 val_407 2010-03-03 22 -421 val_421 2010-03-03 22 -401 val_401 2010-03-03 22 -375 val_375 2010-03-03 22 -454 val_454 2010-03-03 22 -189 val_189 2010-03-03 22 -175 val_175 2010-03-03 22 -133 val_133 2010-03-03 22 406 val_406 2010-03-03 22 -233 val_233 2010-03-03 22 -462 val_462 2010-03-03 22 -214 val_214 2010-03-03 22 -172 val_172 2010-03-03 22 -353 val_353 2010-03-03 22 -136 val_136 2010-03-03 22 -83 val_83 2010-03-03 22 -480 val_480 2010-03-03 22 -265 val_265 2010-03-03 22 -249 val_249 2010-03-03 22 -207 val_207 2010-03-03 22 -460 val_460 2010-03-03 22 -493 val_493 2010-03-03 22 -333 val_333 2010-03-03 22 -317 val_317 2010-03-03 22 -310 val_310 2010-03-03 22 -468 val_468 2010-03-03 22 -178 val_178 2010-03-03 22 -478 val_478 2010-03-03 22 -230 val_230 2010-03-03 22 -277 val_277 2010-03-03 22 -325 val_325 2010-03-03 22 -323 val_323 2010-03-03 22 -443 val_443 2010-03-03 22 -169 val_169 2010-03-03 22 429 val_429 2010-03-03 22 -120 val_120 2010-03-03 22 -444 val_444 2010-03-03 22 -199 val_199 2010-03-03 22 +374 val_374 2010-03-03 22 +152 val_152 2010-03-03 22 +469 val_469 2010-03-03 22 +145 val_145 2010-03-03 22 +495 val_495 2010-03-03 22 +37 val_37 2010-03-03 22 +327 val_327 2010-03-03 22 +281 val_281 2010-03-03 22 +277 val_277 2010-03-03 22 +209 val_209 2010-03-03 22 +15 val_15 2010-03-03 22 +82 val_82 2010-03-03 22 +403 val_403 2010-03-03 22 +166 val_166 2010-03-03 22 417 val_417 2010-03-03 22 -305 val_305 2010-03-03 22 -479 val_479 2010-03-03 22 -248 val_248 2010-03-03 22 -360 val_360 2010-03-03 22 -439 val_439 2010-03-03 22 +430 val_430 2010-03-03 22 +252 val_252 2010-03-03 22 +292 val_292 2010-03-03 22 +219 val_219 2010-03-03 22 +287 val_287 2010-03-03 22 +153 val_153 2010-03-03 22 +193 val_193 2010-03-03 22 +338 val_338 2010-03-03 22 +446 val_446 2010-03-03 22 +459 val_459 2010-03-03 22 +394 val_394 2010-03-03 22 237 val_237 2010-03-03 22 -491 val_491 2010-03-03 22 -200 val_200 2010-03-03 22 -414 val_414 2010-03-03 22 -119 val_119 2010-03-03 22 -438 val_438 2010-03-03 22 -163 val_163 2010-03-03 22 -70 val_70 2010-03-03 22 -104 val_104 2010-03-03 22 -255 val_255 2010-03-03 22 -351 val_351 2010-03-03 22 -24 val_24 2010-03-03 22 -291 val_291 2010-03-03 22 -480 val_480 2010-03-03 22 -397 val_397 2010-03-03 22 -70 val_70 2010-03-03 22 -5 val_5 2010-03-03 22 -382 val_382 2010-03-03 22 -187 val_187 2010-03-03 22 -424 val_424 2010-03-03 22 -164 val_164 2010-03-03 22 -431 val_431 2010-03-03 22 -125 val_125 2010-03-03 22 -298 val_298 2010-03-03 22 -478 val_478 2010-03-03 22 -454 val_454 2010-03-03 22 -431 val_431 2010-03-03 22 -164 val_164 2010-03-03 22 -217 val_217 2010-03-03 22 -201 val_201 2010-03-03 22 +482 val_482 2010-03-03 22 +174 val_174 2010-03-03 22 +413 val_413 2010-03-03 22 +494 val_494 2010-03-03 22 +207 val_207 2010-03-03 22 +199 val_199 2010-03-03 22 +466 val_466 2010-03-03 22 +208 val_208 2010-03-03 22 +174 val_174 2010-03-03 22 +399 val_399 2010-03-03 22 396 val_396 2010-03-03 22 -12 val_12 2010-03-03 22 -424 val_424 2010-03-03 22 -348 val_348 2010-03-03 22 -262 val_262 2010-03-03 22 +247 val_247 2010-03-03 22 +417 val_417 2010-03-03 22 +489 val_489 2010-03-03 22 +162 val_162 2010-03-03 22 +377 val_377 2010-03-03 22 +397 val_397 2010-03-03 22 +309 val_309 2010-03-03 22 +365 val_365 2010-03-03 22 +266 val_266 2010-03-03 22 +439 val_439 2010-03-03 22 +342 val_342 2010-03-03 22 +367 val_367 2010-03-03 22 +325 val_325 2010-03-03 22 +167 val_167 2010-03-03 22 +195 val_195 2010-03-03 22 +475 val_475 2010-03-03 22 +17 val_17 2010-03-03 22 +113 val_113 2010-03-03 22 +155 val_155 2010-03-03 22 203 val_203 2010-03-03 22 -90 val_90 2010-03-03 22 -258 val_258 2010-03-03 22 -114 val_114 2010-03-03 22 -401 val_401 2010-03-03 22 -406 val_406 2010-03-03 22 -190 val_190 2010-03-03 22 -409 val_409 2010-03-03 22 -406 val_406 2010-03-03 22 -257 val_257 2010-03-03 22 -105 val_105 2010-03-03 22 -53 val_53 2010-03-03 22 -483 val_483 2010-03-03 22 -403 val_403 2010-03-03 22 -175 val_175 2010-03-03 22 -366 val_366 2010-03-03 22 -466 val_466 2010-03-03 22 -104 val_104 2010-03-03 22 -335 val_335 2010-03-03 22 -321 val_321 2010-03-03 22 -193 val_193 2010-03-03 22 -44 val_44 2010-03-03 22 -80 val_80 2010-03-03 22 -235 val_235 2010-03-03 22 -331 val_331 2010-03-03 22 -283 val_283 2010-03-03 22 -35 val_35 2010-03-03 22 -2 val_2 2010-03-03 22 -280 val_280 2010-03-03 22 -463 val_463 2010-03-03 22 -469 val_469 2010-03-03 22 -229 val_229 2010-03-03 22 -316 val_316 2010-03-03 22 -202 val_202 2010-03-03 22 -432 val_432 2010-03-03 22 -467 val_467 2010-03-03 22 +339 val_339 2010-03-03 22 +0 val_0 2010-03-03 22 +455 val_455 2010-03-03 22 128 val_128 2010-03-03 22 +311 val_311 2010-03-03 22 +316 val_316 2010-03-03 22 +57 val_57 2010-03-03 22 +302 val_302 2010-03-03 22 +205 val_205 2010-03-03 22 +149 val_149 2010-03-03 22 438 val_438 2010-03-03 22 -244 val_244 2010-03-03 22 -5 val_5 2010-03-03 22 -191 val_191 2010-03-03 22 -288 val_288 2010-03-03 22 -401 val_401 2010-03-03 22 -480 val_480 2010-03-03 22 -487 val_487 2010-03-03 22 -70 val_70 2010-03-03 22 -263 val_263 2010-03-03 22 -256 val_256 2010-03-03 22 -223 val_223 2010-03-03 22 -116 val_116 2010-03-03 22 -485 val_485 2010-03-03 22 +345 val_345 2010-03-03 22 +129 val_129 2010-03-03 22 +170 val_170 2010-03-03 22 +20 val_20 2010-03-03 22 +489 val_489 2010-03-03 22 +157 val_157 2010-03-03 22 +378 val_378 2010-03-03 22 +221 val_221 2010-03-03 22 +92 val_92 2010-03-03 22 +111 val_111 2010-03-03 22 +47 val_47 2010-03-03 22 +72 val_72 2010-03-03 22 +4 val_4 2010-03-03 22 +280 val_280 2010-03-03 22 +35 val_35 2010-03-03 22 +427 val_427 2010-03-03 22 +277 val_277 2010-03-03 22 +208 val_208 2010-03-03 22 +356 val_356 2010-03-03 22 +399 val_399 2010-03-03 22 +169 val_169 2010-03-03 22 +382 val_382 2010-03-03 22 +498 val_498 2010-03-03 22 +125 val_125 2010-03-03 22 +386 val_386 2010-03-03 22 +437 val_437 2010-03-03 22 +469 val_469 2010-03-03 22 +192 val_192 2010-03-03 22 +286 val_286 2010-03-03 22 +187 val_187 2010-03-03 22 +176 val_176 2010-03-03 22 +54 val_54 2010-03-03 22 +459 val_459 2010-03-03 22 +51 val_51 2010-03-03 22 +138 val_138 2010-03-03 22 +103 val_103 2010-03-03 22 239 val_239 2010-03-03 22 -219 val_219 2010-03-03 22 -274 val_274 2010-03-03 22 -167 val_167 2010-03-03 22 -344 val_344 2010-03-03 22 -367 val_367 2010-03-03 22 +213 val_213 2010-03-03 22 216 val_216 2010-03-03 22 -113 val_113 2010-03-03 22 -296 val_296 2010-03-03 22 -103 val_103 2010-03-03 22 -368 val_368 2010-03-03 22 -33 val_33 2010-03-03 22 +430 val_430 2010-03-03 22 +278 val_278 2010-03-03 22 +176 val_176 2010-03-03 22 +289 val_289 2010-03-03 22 +221 val_221 2010-03-03 22 +65 val_65 2010-03-03 22 +318 val_318 2010-03-03 22 +332 val_332 2010-03-03 22 +311 val_311 2010-03-03 22 +275 val_275 2010-03-03 22 +137 val_137 2010-03-03 22 +241 val_241 2010-03-03 22 +83 val_83 2010-03-03 22 +333 val_333 2010-03-03 22 +180 val_180 2010-03-03 22 +284 val_284 2010-03-03 22 +12 val_12 2010-03-03 22 230 val_230 2010-03-03 22 -69 val_69 2010-03-03 22 -342 val_342 2010-03-03 22 -74 val_74 2010-03-03 22 -76 val_76 2010-03-03 22 -468 val_468 2010-03-03 22 -64 val_64 2010-03-03 22 -209 val_209 2010-03-03 22 -30 val_30 2010-03-03 22 -453 val_453 2010-03-03 22 -138 val_138 2010-03-03 22 -228 val_228 2010-03-03 22 -218 val_218 2010-03-03 22 -449 val_449 2010-03-03 22 -149 val_149 2010-03-03 22 -492 val_492 2010-03-03 22 -223 val_223 2010-03-03 22 -41 val_41 2010-03-03 22 -76 val_76 2010-03-03 22 -78 val_78 2010-03-03 22 -458 val_458 2010-03-03 22 +181 val_181 2010-03-03 22 +67 val_67 2010-03-03 22 +260 val_260 2010-03-03 22 +404 val_404 2010-03-03 22 +384 val_384 2010-03-03 22 489 val_489 2010-03-03 22 -119 val_119 2010-03-03 22 -430 val_430 2010-03-03 22 -321 val_321 2010-03-03 22 +353 val_353 2010-03-03 22 +373 val_373 2010-03-03 22 +272 val_272 2010-03-03 22 +138 val_138 2010-03-03 22 +217 val_217 2010-03-03 22 +84 val_84 2010-03-03 22 +348 val_348 2010-03-03 22 +466 val_466 2010-03-03 22 +58 val_58 2010-03-03 22 +8 val_8 2010-03-03 22 +411 val_411 2010-03-03 22 +230 val_230 2010-03-03 22 +208 val_208 2010-03-03 22 +348 val_348 2010-03-03 22 +24 val_24 2010-03-03 22 +463 val_463 2010-03-03 22 +431 val_431 2010-03-03 22 +179 val_179 2010-03-03 22 +172 val_172 2010-03-03 22 42 val_42 2010-03-03 22 -195 val_195 2010-03-03 22 -160 val_160 2010-03-03 22 -498 val_498 2010-03-03 22 +129 val_129 2010-03-03 22 +158 val_158 2010-03-03 22 +119 val_119 2010-03-03 22 +496 val_496 2010-03-03 22 +0 val_0 2010-03-03 22 322 val_322 2010-03-03 22 -472 val_472 2010-03-03 22 -143 val_143 2010-03-03 22 -233 val_233 2010-03-03 22 -229 val_229 2010-03-03 22 -34 val_34 2010-03-03 22 -168 val_168 2010-03-03 22 -11 val_11 2010-03-03 22 -95 val_95 2010-03-03 22 -336 val_336 2010-03-03 22 -35 val_35 2010-03-03 22 -58 val_58 2010-03-03 22 -395 val_395 2010-03-03 22 -317 val_317 2010-03-03 22 -396 val_396 2010-03-03 22 -402 val_402 2010-03-03 22 -497 val_497 2010-03-03 22 -5 val_5 2010-03-03 22 -226 val_226 2010-03-03 22 -177 val_177 2010-03-03 22 -452 val_452 2010-03-03 22 -242 val_242 2010-03-03 22 -401 val_401 2010-03-03 22 -331 val_331 2010-03-03 22 -272 val_272 2010-03-03 22 -392 val_392 2010-03-03 22 -369 val_369 2010-03-03 22 -242 val_242 2010-03-03 22 -327 val_327 2010-03-03 22 -389 val_389 2010-03-03 22 -309 val_309 2010-03-03 22 -224 val_224 2010-03-03 22 -306 val_306 2010-03-03 22 -273 val_273 2010-03-03 22 -277 val_277 2010-03-03 22 -10 val_10 2010-03-03 22 -435 val_435 2010-03-03 22 -19 val_19 2010-03-03 22 -307 val_307 2010-03-03 22 -90 val_90 2010-03-03 22 -72 val_72 2010-03-03 22 -118 val_118 2010-03-03 22 -15 val_15 2010-03-03 22 -419 val_419 2010-03-03 22 -238 val_238 2010-03-03 22 -138 val_138 2010-03-03 22 -282 val_282 2010-03-03 22 -395 val_395 2010-03-03 22 -134 val_134 2010-03-03 22 -118 val_118 2010-03-03 22 -179 val_179 2010-03-03 22 -364 val_364 2010-03-03 22 -87 val_87 2010-03-03 22 -490 val_490 2010-03-03 22 -0 val_0 2010-03-03 22 -77 val_77 2010-03-03 22 -85 val_85 2010-03-03 22 -413 val_413 2010-03-03 22 -169 val_169 2010-03-03 22 -316 val_316 2010-03-03 22 -369 val_369 2010-03-03 22 -137 val_137 2010-03-03 22 -470 val_470 2010-03-03 22 -409 val_409 2010-03-03 22 -318 val_318 2010-03-03 22 -318 val_318 2010-03-03 22 -187 val_187 2010-03-03 22 197 val_197 2010-03-03 22 -282 val_282 2010-03-03 22 -98 val_98 2010-03-03 22 -457 val_457 2010-03-03 22 -481 val_481 2010-03-03 22 -288 val_288 2010-03-03 22 -196 val_196 2010-03-03 22 -95 val_95 2010-03-03 22 -308 val_308 2010-03-03 22 468 val_468 2010-03-03 22 -469 val_469 2010-03-03 22 -156 val_156 2010-03-03 22 -436 val_436 2010-03-03 22 -43 val_43 2010-03-03 22 -404 val_404 2010-03-03 22 -51 val_51 2010-03-03 22 -131 val_131 2010-03-03 22 -120 val_120 2010-03-03 22 -205 val_205 2010-03-03 22 -230 val_230 2010-03-03 22 -327 val_327 2010-03-03 22 -165 val_165 2010-03-03 22 -26 val_26 2010-03-03 22 -96 val_96 2010-03-03 22 -418 val_418 2010-03-03 22 -191 val_191 2010-03-03 22 -199 val_199 2010-03-03 22 -298 val_298 2010-03-03 22 -100 val_100 2010-03-03 22 -454 val_454 2010-03-03 22 393 val_393 2010-03-03 22 +454 val_454 2010-03-03 22 +100 val_100 2010-03-03 22 +298 val_298 2010-03-03 22 +199 val_199 2010-03-03 22 +191 val_191 2010-03-03 22 +418 val_418 2010-03-03 22 +96 val_96 2010-03-03 22 +26 val_26 2010-03-03 22 +165 val_165 2010-03-03 22 +327 val_327 2010-03-03 22 +230 val_230 2010-03-03 22 +205 val_205 2010-03-03 22 +120 val_120 2010-03-03 22 +131 val_131 2010-03-03 22 +51 val_51 2010-03-03 22 +404 val_404 2010-03-03 22 +43 val_43 2010-03-03 22 +436 val_436 2010-03-03 22 +156 val_156 2010-03-03 22 +469 val_469 2010-03-03 22 468 val_468 2010-03-03 22 +308 val_308 2010-03-03 22 +95 val_95 2010-03-03 22 +196 val_196 2010-03-03 22 +288 val_288 2010-03-03 22 +481 val_481 2010-03-03 22 +457 val_457 2010-03-03 22 +98 val_98 2010-03-03 22 +282 val_282 2010-03-03 22 197 val_197 2010-03-03 22 -322 val_322 2010-03-03 22 +187 val_187 2010-03-03 22 +318 val_318 2010-03-03 22 +318 val_318 2010-03-03 22 +409 val_409 2010-03-03 22 +470 val_470 2010-03-03 22 +137 val_137 2010-03-03 22 +369 val_369 2010-03-03 22 +316 val_316 2010-03-03 22 +169 val_169 2010-03-03 22 +413 val_413 2010-03-03 22 +85 val_85 2010-03-03 22 +77 val_77 2010-03-03 22 0 val_0 2010-03-03 22 -496 val_496 2010-03-03 22 -119 val_119 2010-03-03 22 -158 val_158 2010-03-03 22 -129 val_129 2010-03-03 22 -42 val_42 2010-03-03 22 -172 val_172 2010-03-03 22 +490 val_490 2010-03-03 22 +87 val_87 2010-03-03 22 +364 val_364 2010-03-03 22 179 val_179 2010-03-03 22 -431 val_431 2010-03-03 22 -463 val_463 2010-03-03 22 -24 val_24 2010-03-03 22 -348 val_348 2010-03-03 22 -208 val_208 2010-03-03 22 -230 val_230 2010-03-03 22 -411 val_411 2010-03-03 22 -8 val_8 2010-03-03 22 -58 val_58 2010-03-03 22 -466 val_466 2010-03-03 22 -348 val_348 2010-03-03 22 -84 val_84 2010-03-03 22 -217 val_217 2010-03-03 22 +118 val_118 2010-03-03 22 +134 val_134 2010-03-03 22 +395 val_395 2010-03-03 22 +282 val_282 2010-03-03 22 138 val_138 2010-03-03 22 +238 val_238 2010-03-03 22 +419 val_419 2010-03-03 22 +15 val_15 2010-03-03 22 +118 val_118 2010-03-03 22 +72 val_72 2010-03-03 22 +90 val_90 2010-03-03 22 +307 val_307 2010-03-03 22 +19 val_19 2010-03-03 22 +435 val_435 2010-03-03 22 +10 val_10 2010-03-03 22 +277 val_277 2010-03-03 22 +273 val_273 2010-03-03 22 +306 val_306 2010-03-03 22 +224 val_224 2010-03-03 22 +309 val_309 2010-03-03 22 +389 val_389 2010-03-03 22 +327 val_327 2010-03-03 22 +242 val_242 2010-03-03 22 +369 val_369 2010-03-03 22 +392 val_392 2010-03-03 22 272 val_272 2010-03-03 22 -373 val_373 2010-03-03 22 -353 val_353 2010-03-03 22 +331 val_331 2010-03-03 22 +401 val_401 2010-03-03 22 +242 val_242 2010-03-03 22 +452 val_452 2010-03-03 22 +177 val_177 2010-03-03 22 +226 val_226 2010-03-03 22 +5 val_5 2010-03-03 22 +497 val_497 2010-03-03 22 +402 val_402 2010-03-03 22 +396 val_396 2010-03-03 22 +317 val_317 2010-03-03 22 +395 val_395 2010-03-03 22 +58 val_58 2010-03-03 22 +35 val_35 2010-03-03 22 +336 val_336 2010-03-03 22 +95 val_95 2010-03-03 22 +11 val_11 2010-03-03 22 +168 val_168 2010-03-03 22 +34 val_34 2010-03-03 22 +229 val_229 2010-03-03 22 +233 val_233 2010-03-03 22 +143 val_143 2010-03-03 22 +472 val_472 2010-03-03 22 +322 val_322 2010-03-03 22 +498 val_498 2010-03-03 22 +160 val_160 2010-03-03 22 +195 val_195 2010-03-03 22 +42 val_42 2010-03-03 22 +321 val_321 2010-03-03 22 +430 val_430 2010-03-03 22 +119 val_119 2010-03-03 22 489 val_489 2010-03-03 22 -384 val_384 2010-03-03 22 -404 val_404 2010-03-03 22 -260 val_260 2010-03-03 22 -67 val_67 2010-03-03 22 -181 val_181 2010-03-03 22 +458 val_458 2010-03-03 22 +78 val_78 2010-03-03 22 +76 val_76 2010-03-03 22 +41 val_41 2010-03-03 22 +223 val_223 2010-03-03 22 +492 val_492 2010-03-03 22 +149 val_149 2010-03-03 22 +449 val_449 2010-03-03 22 +218 val_218 2010-03-03 22 +228 val_228 2010-03-03 22 +138 val_138 2010-03-03 22 +453 val_453 2010-03-03 22 +30 val_30 2010-03-03 22 +209 val_209 2010-03-03 22 +64 val_64 2010-03-03 22 +468 val_468 2010-03-03 22 +76 val_76 2010-03-03 22 +74 val_74 2010-03-03 22 +342 val_342 2010-03-03 22 +69 val_69 2010-03-03 22 230 val_230 2010-03-03 22 -12 val_12 2010-03-03 22 -284 val_284 2010-03-03 22 -180 val_180 2010-03-03 22 -333 val_333 2010-03-03 22 -83 val_83 2010-03-03 22 -241 val_241 2010-03-03 22 -137 val_137 2010-03-03 22 -275 val_275 2010-03-03 22 -311 val_311 2010-03-03 22 -332 val_332 2010-03-03 22 -318 val_318 2010-03-03 22 -65 val_65 2010-03-03 22 -221 val_221 2010-03-03 22 -289 val_289 2010-03-03 22 -176 val_176 2010-03-03 22 -278 val_278 2010-03-03 22 -430 val_430 2010-03-03 22 -216 val_216 2010-03-03 22 -213 val_213 2010-03-03 22 -239 val_239 2010-03-03 22 +33 val_33 2010-03-03 22 +368 val_368 2010-03-03 22 103 val_103 2010-03-03 22 -138 val_138 2010-03-03 22 -51 val_51 2010-03-03 22 -459 val_459 2010-03-03 22 -54 val_54 2010-03-03 22 -176 val_176 2010-03-03 22 -187 val_187 2010-03-03 22 -286 val_286 2010-03-03 22 -192 val_192 2010-03-03 22 +296 val_296 2010-03-03 22 +113 val_113 2010-03-03 22 +216 val_216 2010-03-03 22 +367 val_367 2010-03-03 22 +344 val_344 2010-03-03 22 +167 val_167 2010-03-03 22 +274 val_274 2010-03-03 22 +219 val_219 2010-03-03 22 +239 val_239 2010-03-03 22 +485 val_485 2010-03-03 22 +116 val_116 2010-03-03 22 +223 val_223 2010-03-03 22 +256 val_256 2010-03-03 22 +263 val_263 2010-03-03 22 +70 val_70 2010-03-03 22 +487 val_487 2010-03-03 22 +480 val_480 2010-03-03 22 +401 val_401 2010-03-03 22 +288 val_288 2010-03-03 22 +191 val_191 2010-03-03 22 +5 val_5 2010-03-03 22 +244 val_244 2010-03-03 22 +438 val_438 2010-03-03 22 +128 val_128 2010-03-03 22 +467 val_467 2010-03-03 22 +432 val_432 2010-03-03 22 +202 val_202 2010-03-03 22 +316 val_316 2010-03-03 22 +229 val_229 2010-03-03 22 469 val_469 2010-03-03 22 -437 val_437 2010-03-03 22 -386 val_386 2010-03-03 22 +463 val_463 2010-03-03 22 +280 val_280 2010-03-03 22 +2 val_2 2010-03-03 22 +35 val_35 2010-03-03 22 +283 val_283 2010-03-03 22 +331 val_331 2010-03-03 22 +235 val_235 2010-03-03 22 +80 val_80 2010-03-03 22 +44 val_44 2010-03-03 22 +193 val_193 2010-03-03 22 +321 val_321 2010-03-03 22 +335 val_335 2010-03-03 22 +104 val_104 2010-03-03 22 +466 val_466 2010-03-03 22 +366 val_366 2010-03-03 22 +175 val_175 2010-03-03 22 +403 val_403 2010-03-03 22 +483 val_483 2010-03-03 22 +53 val_53 2010-03-03 22 +105 val_105 2010-03-03 22 +257 val_257 2010-03-03 22 +406 val_406 2010-03-03 22 +409 val_409 2010-03-03 22 +190 val_190 2010-03-03 22 +406 val_406 2010-03-03 22 +401 val_401 2010-03-03 22 +114 val_114 2010-03-03 22 +258 val_258 2010-03-03 22 +90 val_90 2010-03-03 22 +203 val_203 2010-03-03 22 +262 val_262 2010-03-03 22 +348 val_348 2010-03-03 22 +424 val_424 2010-03-03 22 +12 val_12 2010-03-03 22 +396 val_396 2010-03-03 22 +201 val_201 2010-03-03 22 +217 val_217 2010-03-03 22 +164 val_164 2010-03-03 22 +431 val_431 2010-03-03 22 +454 val_454 2010-03-03 22 +478 val_478 2010-03-03 22 +298 val_298 2010-03-03 22 125 val_125 2010-03-03 22 -498 val_498 2010-03-03 22 +431 val_431 2010-03-03 22 +164 val_164 2010-03-03 22 +424 val_424 2010-03-03 22 +187 val_187 2010-03-03 22 382 val_382 2010-03-03 22 -169 val_169 2010-03-03 22 -399 val_399 2010-03-03 22 -356 val_356 2010-03-03 22 -208 val_208 2010-03-03 22 -277 val_277 2010-03-03 22 -427 val_427 2010-03-03 22 -35 val_35 2010-03-03 22 -280 val_280 2010-03-03 22 -4 val_4 2010-03-03 22 -72 val_72 2010-03-03 22 -47 val_47 2010-03-03 22 -111 val_111 2010-03-03 22 -92 val_92 2010-03-03 22 -221 val_221 2010-03-03 22 -378 val_378 2010-03-03 22 -157 val_157 2010-03-03 22 -489 val_489 2010-03-03 22 -20 val_20 2010-03-03 22 -170 val_170 2010-03-03 22 -129 val_129 2010-03-03 22 -345 val_345 2010-03-03 22 +5 val_5 2010-03-03 22 +70 val_70 2010-03-03 22 +397 val_397 2010-03-03 22 +480 val_480 2010-03-03 22 +291 val_291 2010-03-03 22 +24 val_24 2010-03-03 22 +351 val_351 2010-03-03 22 +255 val_255 2010-03-03 22 +104 val_104 2010-03-03 22 +70 val_70 2010-03-03 22 +163 val_163 2010-03-03 22 438 val_438 2010-03-03 22 -149 val_149 2010-03-03 22 -205 val_205 2010-03-03 22 -302 val_302 2010-03-03 22 -57 val_57 2010-03-03 22 -316 val_316 2010-03-03 22 -311 val_311 2010-03-03 22 -128 val_128 2010-03-03 22 -455 val_455 2010-03-03 22 -0 val_0 2010-03-03 22 -339 val_339 2010-03-03 22 -203 val_203 2010-03-03 22 -155 val_155 2010-03-03 22 -113 val_113 2010-03-03 22 -17 val_17 2010-03-03 22 -475 val_475 2010-03-03 22 -195 val_195 2010-03-03 22 -167 val_167 2010-03-03 22 -325 val_325 2010-03-03 22 -367 val_367 2010-03-03 22 -342 val_342 2010-03-03 22 +119 val_119 2010-03-03 22 +414 val_414 2010-03-03 22 +200 val_200 2010-03-03 22 +491 val_491 2010-03-03 22 +237 val_237 2010-03-03 22 439 val_439 2010-03-03 22 -266 val_266 2010-03-03 22 -365 val_365 2010-03-03 22 -309 val_309 2010-03-03 22 -397 val_397 2010-03-03 22 -377 val_377 2010-03-03 22 -162 val_162 2010-03-03 22 -489 val_489 2010-03-03 22 +360 val_360 2010-03-03 22 +248 val_248 2010-03-03 22 +479 val_479 2010-03-03 22 +305 val_305 2010-03-03 22 417 val_417 2010-03-03 22 -247 val_247 2010-03-03 22 -396 val_396 2010-03-03 22 -399 val_399 2010-03-03 22 -174 val_174 2010-03-03 22 -208 val_208 2010-03-03 22 -466 val_466 2010-03-03 22 199 val_199 2010-03-03 22 -207 val_207 2010-03-03 22 -494 val_494 2010-03-03 22 -413 val_413 2010-03-03 22 -174 val_174 2010-03-03 22 -482 val_482 2010-03-03 22 -237 val_237 2010-03-03 22 -394 val_394 2010-03-03 22 -459 val_459 2010-03-03 22 -446 val_446 2010-03-03 22 -338 val_338 2010-03-03 22 -193 val_193 2010-03-03 22 -153 val_153 2010-03-03 22 -287 val_287 2010-03-03 22 -219 val_219 2010-03-03 22 -292 val_292 2010-03-03 22 -252 val_252 2010-03-03 22 -430 val_430 2010-03-03 22 -417 val_417 2010-03-03 22 -166 val_166 2010-03-03 22 -403 val_403 2010-03-03 22 -82 val_82 2010-03-03 22 -15 val_15 2010-03-03 22 -209 val_209 2010-03-03 22 -277 val_277 2010-03-03 22 -281 val_281 2010-03-03 22 -327 val_327 2010-03-03 22 -37 val_37 2010-03-03 22 -495 val_495 2010-03-03 22 -145 val_145 2010-03-03 22 -469 val_469 2010-03-03 22 -152 val_152 2010-03-03 22 -374 val_374 2010-03-03 22 +444 val_444 2010-03-03 22 +120 val_120 2010-03-03 22 429 val_429 2010-03-03 22 +169 val_169 2010-03-03 22 +443 val_443 2010-03-03 22 +323 val_323 2010-03-03 22 +325 val_325 2010-03-03 22 +277 val_277 2010-03-03 22 +230 val_230 2010-03-03 22 +478 val_478 2010-03-03 22 +178 val_178 2010-03-03 22 +468 val_468 2010-03-03 22 +310 val_310 2010-03-03 22 +317 val_317 2010-03-03 22 +333 val_333 2010-03-03 22 +493 val_493 2010-03-03 22 +460 val_460 2010-03-03 22 +207 val_207 2010-03-03 22 +249 val_249 2010-03-03 22 +265 val_265 2010-03-03 22 +480 val_480 2010-03-03 22 +83 val_83 2010-03-03 22 +136 val_136 2010-03-03 22 +353 val_353 2010-03-03 22 +172 val_172 2010-03-03 22 +214 val_214 2010-03-03 22 +462 val_462 2010-03-03 22 +233 val_233 2010-03-03 22 406 val_406 2010-03-03 22 +133 val_133 2010-03-03 22 +175 val_175 2010-03-03 22 +189 val_189 2010-03-03 22 +454 val_454 2010-03-03 22 +375 val_375 2010-03-03 22 +401 val_401 2010-03-03 22 +421 val_421 2010-03-03 22 +407 val_407 2010-03-03 22 +384 val_384 2010-03-03 22 +256 val_256 2010-03-03 22 +26 val_26 2010-03-03 22 +134 val_134 2010-03-03 22 +67 val_67 2010-03-03 22 +384 val_384 2010-03-03 22 +379 val_379 2010-03-03 22 +18 val_18 2010-03-03 22 +462 val_462 2010-03-03 22 +492 val_492 2010-03-03 22 +100 val_100 2010-03-03 22 +298 val_298 2010-03-03 22 +9 val_9 2010-03-03 22 +341 val_341 2010-03-03 22 +498 val_498 2010-03-03 22 146 val_146 2010-03-03 22 -213 val_213 2010-03-03 22 -128 val_128 2010-03-03 22 -66 val_66 2010-03-03 22 -369 val_369 2010-03-03 22 -224 val_224 2010-03-03 22 +458 val_458 2010-03-03 22 +362 val_362 2010-03-03 22 +186 val_186 2010-03-03 22 +285 val_285 2010-03-03 22 +348 val_348 2010-03-03 22 +167 val_167 2010-03-03 22 +18 val_18 2010-03-03 22 273 val_273 2010-03-03 22 -150 val_150 2010-03-03 22 -401 val_401 2010-03-03 22 -193 val_193 2010-03-03 22 -265 val_265 2010-03-03 22 -484 val_484 2010-03-03 22 -98 val_98 2010-03-03 22 -278 val_278 2010-03-03 22 -255 val_255 2010-03-03 22 -409 val_409 2010-03-03 22 -165 val_165 2010-03-03 22 -27 val_27 2010-03-03 22 -311 val_311 2010-03-03 22 -86 val_86 2010-03-03 22 -238 val_238 2010-03-03 22 -97 val_97 2010-03-03 24 -200 val_200 2010-03-03 24 -400 val_400 2010-03-03 24 -403 val_403 2010-03-03 24 -169 val_169 2010-03-03 24 -90 val_90 2010-03-03 24 -126 val_126 2010-03-03 24 -222 val_222 2010-03-03 24 -477 val_477 2010-03-03 24 -414 val_414 2010-03-03 24 -194 val_194 2010-03-03 24 -307 val_307 2010-03-03 24 -348 val_348 2010-03-03 24 -152 val_152 2010-03-03 24 -448 val_448 2010-03-03 24 -37 val_37 2010-03-03 24 -28 val_28 2010-03-03 24 -84 val_84 2010-03-03 24 -315 val_315 2010-03-03 24 -469 val_469 2010-03-03 24 -97 val_97 2010-03-03 24 -344 val_344 2010-03-03 24 -281 val_281 2010-03-03 24 -183 val_183 2010-03-03 24 +183 val_183 2010-03-03 22 +281 val_281 2010-03-03 22 +344 val_344 2010-03-03 22 +97 val_97 2010-03-03 22 +469 val_469 2010-03-03 22 +315 val_315 2010-03-03 22 +84 val_84 2010-03-03 22 +28 val_28 2010-03-03 22 +37 val_37 2010-03-03 22 +448 val_448 2010-03-03 22 +152 val_152 2010-03-03 22 +348 val_348 2010-03-03 22 +307 val_307 2010-03-03 22 +194 val_194 2010-03-03 22 +414 val_414 2010-03-03 22 +477 val_477 2010-03-03 22 +222 val_222 2010-03-03 22 +126 val_126 2010-03-03 22 +90 val_90 2010-03-03 22 +169 val_169 2010-03-03 22 +403 val_403 2010-03-03 22 +400 val_400 2010-03-03 22 +200 val_200 2010-03-03 22 +97 val_97 2010-03-03 22 +238 val_238 2010-03-03 24 +86 val_86 2010-03-03 24 +311 val_311 2010-03-03 24 +27 val_27 2010-03-03 24 +165 val_165 2010-03-03 24 +409 val_409 2010-03-03 24 +255 val_255 2010-03-03 24 +278 val_278 2010-03-03 24 +98 val_98 2010-03-03 24 +484 val_484 2010-03-03 24 +265 val_265 2010-03-03 24 +193 val_193 2010-03-03 24 +401 val_401 2010-03-03 24 +150 val_150 2010-03-03 24 273 val_273 2010-03-03 24 -18 val_18 2010-03-03 24 -167 val_167 2010-03-03 24 -348 val_348 2010-03-03 24 -285 val_285 2010-03-03 24 -186 val_186 2010-03-03 24 -362 val_362 2010-03-03 24 -458 val_458 2010-03-03 24 +224 val_224 2010-03-03 24 +369 val_369 2010-03-03 24 +66 val_66 2010-03-03 24 +128 val_128 2010-03-03 24 +213 val_213 2010-03-03 24 146 val_146 2010-03-03 24 -498 val_498 2010-03-03 24 -341 val_341 2010-03-03 24 -9 val_9 2010-03-03 24 -298 val_298 2010-03-03 24 -100 val_100 2010-03-03 24 -492 val_492 2010-03-03 24 -462 val_462 2010-03-03 24 -18 val_18 2010-03-03 24 -379 val_379 2010-03-03 24 -384 val_384 2010-03-03 24 -67 val_67 2010-03-03 24 -134 val_134 2010-03-03 24 -26 val_26 2010-03-03 24 -256 val_256 2010-03-03 24 -384 val_384 2010-03-03 24 -407 val_407 2010-03-03 24 -421 val_421 2010-03-03 24 -401 val_401 2010-03-03 24 -375 val_375 2010-03-03 24 -454 val_454 2010-03-03 24 -189 val_189 2010-03-03 24 -175 val_175 2010-03-03 24 -133 val_133 2010-03-03 24 406 val_406 2010-03-03 24 -233 val_233 2010-03-03 24 -462 val_462 2010-03-03 24 -214 val_214 2010-03-03 24 -172 val_172 2010-03-03 24 -353 val_353 2010-03-03 24 -136 val_136 2010-03-03 24 -83 val_83 2010-03-03 24 -480 val_480 2010-03-03 24 -265 val_265 2010-03-03 24 -249 val_249 2010-03-03 24 -207 val_207 2010-03-03 24 -460 val_460 2010-03-03 24 -493 val_493 2010-03-03 24 -333 val_333 2010-03-03 24 -317 val_317 2010-03-03 24 -310 val_310 2010-03-03 24 -468 val_468 2010-03-03 24 -178 val_178 2010-03-03 24 -478 val_478 2010-03-03 24 -230 val_230 2010-03-03 24 -277 val_277 2010-03-03 24 -325 val_325 2010-03-03 24 -323 val_323 2010-03-03 24 -443 val_443 2010-03-03 24 -169 val_169 2010-03-03 24 429 val_429 2010-03-03 24 -120 val_120 2010-03-03 24 -444 val_444 2010-03-03 24 -199 val_199 2010-03-03 24 +374 val_374 2010-03-03 24 +152 val_152 2010-03-03 24 +469 val_469 2010-03-03 24 +145 val_145 2010-03-03 24 +495 val_495 2010-03-03 24 +37 val_37 2010-03-03 24 +327 val_327 2010-03-03 24 +281 val_281 2010-03-03 24 +277 val_277 2010-03-03 24 +209 val_209 2010-03-03 24 +15 val_15 2010-03-03 24 +82 val_82 2010-03-03 24 +403 val_403 2010-03-03 24 +166 val_166 2010-03-03 24 417 val_417 2010-03-03 24 -305 val_305 2010-03-03 24 -479 val_479 2010-03-03 24 -248 val_248 2010-03-03 24 -360 val_360 2010-03-03 24 -439 val_439 2010-03-03 24 +430 val_430 2010-03-03 24 +252 val_252 2010-03-03 24 +292 val_292 2010-03-03 24 +219 val_219 2010-03-03 24 +287 val_287 2010-03-03 24 +153 val_153 2010-03-03 24 +193 val_193 2010-03-03 24 +338 val_338 2010-03-03 24 +446 val_446 2010-03-03 24 +459 val_459 2010-03-03 24 +394 val_394 2010-03-03 24 237 val_237 2010-03-03 24 -491 val_491 2010-03-03 24 -200 val_200 2010-03-03 24 -414 val_414 2010-03-03 24 -119 val_119 2010-03-03 24 -438 val_438 2010-03-03 24 -163 val_163 2010-03-03 24 -70 val_70 2010-03-03 24 -104 val_104 2010-03-03 24 -255 val_255 2010-03-03 24 -351 val_351 2010-03-03 24 -24 val_24 2010-03-03 24 -291 val_291 2010-03-03 24 -480 val_480 2010-03-03 24 -397 val_397 2010-03-03 24 -70 val_70 2010-03-03 24 -5 val_5 2010-03-03 24 -382 val_382 2010-03-03 24 -187 val_187 2010-03-03 24 -424 val_424 2010-03-03 24 -164 val_164 2010-03-03 24 -431 val_431 2010-03-03 24 -125 val_125 2010-03-03 24 -298 val_298 2010-03-03 24 -478 val_478 2010-03-03 24 -454 val_454 2010-03-03 24 -431 val_431 2010-03-03 24 -164 val_164 2010-03-03 24 -217 val_217 2010-03-03 24 -201 val_201 2010-03-03 24 +482 val_482 2010-03-03 24 +174 val_174 2010-03-03 24 +413 val_413 2010-03-03 24 +494 val_494 2010-03-03 24 +207 val_207 2010-03-03 24 +199 val_199 2010-03-03 24 +466 val_466 2010-03-03 24 +208 val_208 2010-03-03 24 +174 val_174 2010-03-03 24 +399 val_399 2010-03-03 24 396 val_396 2010-03-03 24 -12 val_12 2010-03-03 24 -424 val_424 2010-03-03 24 -348 val_348 2010-03-03 24 -262 val_262 2010-03-03 24 +247 val_247 2010-03-03 24 +417 val_417 2010-03-03 24 +489 val_489 2010-03-03 24 +162 val_162 2010-03-03 24 +377 val_377 2010-03-03 24 +397 val_397 2010-03-03 24 +309 val_309 2010-03-03 24 +365 val_365 2010-03-03 24 +266 val_266 2010-03-03 24 +439 val_439 2010-03-03 24 +342 val_342 2010-03-03 24 +367 val_367 2010-03-03 24 +325 val_325 2010-03-03 24 +167 val_167 2010-03-03 24 +195 val_195 2010-03-03 24 +475 val_475 2010-03-03 24 +17 val_17 2010-03-03 24 +113 val_113 2010-03-03 24 +155 val_155 2010-03-03 24 203 val_203 2010-03-03 24 -90 val_90 2010-03-03 24 -258 val_258 2010-03-03 24 -114 val_114 2010-03-03 24 -401 val_401 2010-03-03 24 -406 val_406 2010-03-03 24 -190 val_190 2010-03-03 24 -409 val_409 2010-03-03 24 -406 val_406 2010-03-03 24 -257 val_257 2010-03-03 24 -105 val_105 2010-03-03 24 -53 val_53 2010-03-03 24 -483 val_483 2010-03-03 24 -403 val_403 2010-03-03 24 -175 val_175 2010-03-03 24 -366 val_366 2010-03-03 24 -466 val_466 2010-03-03 24 -104 val_104 2010-03-03 24 -335 val_335 2010-03-03 24 -321 val_321 2010-03-03 24 -193 val_193 2010-03-03 24 -44 val_44 2010-03-03 24 -80 val_80 2010-03-03 24 -235 val_235 2010-03-03 24 -331 val_331 2010-03-03 24 -283 val_283 2010-03-03 24 -35 val_35 2010-03-03 24 -2 val_2 2010-03-03 24 -280 val_280 2010-03-03 24 -463 val_463 2010-03-03 24 -469 val_469 2010-03-03 24 -229 val_229 2010-03-03 24 -316 val_316 2010-03-03 24 -202 val_202 2010-03-03 24 -432 val_432 2010-03-03 24 -467 val_467 2010-03-03 24 +339 val_339 2010-03-03 24 +0 val_0 2010-03-03 24 +455 val_455 2010-03-03 24 128 val_128 2010-03-03 24 +311 val_311 2010-03-03 24 +316 val_316 2010-03-03 24 +57 val_57 2010-03-03 24 +302 val_302 2010-03-03 24 +205 val_205 2010-03-03 24 +149 val_149 2010-03-03 24 438 val_438 2010-03-03 24 -244 val_244 2010-03-03 24 -5 val_5 2010-03-03 24 -191 val_191 2010-03-03 24 -288 val_288 2010-03-03 24 -401 val_401 2010-03-03 24 -480 val_480 2010-03-03 24 -487 val_487 2010-03-03 24 -70 val_70 2010-03-03 24 -263 val_263 2010-03-03 24 -256 val_256 2010-03-03 24 -223 val_223 2010-03-03 24 -116 val_116 2010-03-03 24 -485 val_485 2010-03-03 24 -239 val_239 2010-03-03 24 -219 val_219 2010-03-03 24 -274 val_274 2010-03-03 24 -167 val_167 2010-03-03 24 -344 val_344 2010-03-03 24 -367 val_367 2010-03-03 24 -216 val_216 2010-03-03 24 -113 val_113 2010-03-03 24 -296 val_296 2010-03-03 24 -103 val_103 2010-03-03 24 -368 val_368 2010-03-03 24 -33 val_33 2010-03-03 24 -230 val_230 2010-03-03 24 -69 val_69 2010-03-03 24 -342 val_342 2010-03-03 24 -74 val_74 2010-03-03 24 -76 val_76 2010-03-03 24 -468 val_468 2010-03-03 24 -64 val_64 2010-03-03 24 -209 val_209 2010-03-03 24 -30 val_30 2010-03-03 24 -453 val_453 2010-03-03 24 +345 val_345 2010-03-03 24 +129 val_129 2010-03-03 24 +170 val_170 2010-03-03 24 +20 val_20 2010-03-03 24 +489 val_489 2010-03-03 24 +157 val_157 2010-03-03 24 +378 val_378 2010-03-03 24 +221 val_221 2010-03-03 24 +92 val_92 2010-03-03 24 +111 val_111 2010-03-03 24 +47 val_47 2010-03-03 24 +72 val_72 2010-03-03 24 +4 val_4 2010-03-03 24 +280 val_280 2010-03-03 24 +35 val_35 2010-03-03 24 +427 val_427 2010-03-03 24 +277 val_277 2010-03-03 24 +208 val_208 2010-03-03 24 +356 val_356 2010-03-03 24 +399 val_399 2010-03-03 24 +169 val_169 2010-03-03 24 +382 val_382 2010-03-03 24 +498 val_498 2010-03-03 24 +125 val_125 2010-03-03 24 +386 val_386 2010-03-03 24 +437 val_437 2010-03-03 24 +469 val_469 2010-03-03 24 +192 val_192 2010-03-03 24 +286 val_286 2010-03-03 24 +187 val_187 2010-03-03 24 +176 val_176 2010-03-03 24 +54 val_54 2010-03-03 24 +459 val_459 2010-03-03 24 +51 val_51 2010-03-03 24 138 val_138 2010-03-03 24 -228 val_228 2010-03-03 24 -218 val_218 2010-03-03 24 -449 val_449 2010-03-03 24 -149 val_149 2010-03-03 24 -492 val_492 2010-03-03 24 -223 val_223 2010-03-03 24 -41 val_41 2010-03-03 24 -76 val_76 2010-03-03 24 -78 val_78 2010-03-03 24 -458 val_458 2010-03-03 24 -489 val_489 2010-03-03 24 -119 val_119 2010-03-03 24 +103 val_103 2010-03-03 24 +239 val_239 2010-03-03 24 +213 val_213 2010-03-03 24 +216 val_216 2010-03-03 24 430 val_430 2010-03-03 24 -321 val_321 2010-03-03 24 -42 val_42 2010-03-03 24 -195 val_195 2010-03-03 24 -160 val_160 2010-03-03 24 -498 val_498 2010-03-03 24 -322 val_322 2010-03-03 24 -472 val_472 2010-03-03 24 -143 val_143 2010-03-03 24 -233 val_233 2010-03-03 24 -229 val_229 2010-03-03 24 -34 val_34 2010-03-03 24 -168 val_168 2010-03-03 24 -11 val_11 2010-03-03 24 -95 val_95 2010-03-03 24 -336 val_336 2010-03-03 24 -35 val_35 2010-03-03 24 -58 val_58 2010-03-03 24 -395 val_395 2010-03-03 24 -317 val_317 2010-03-03 24 -396 val_396 2010-03-03 24 -402 val_402 2010-03-03 24 -497 val_497 2010-03-03 24 -5 val_5 2010-03-03 24 -226 val_226 2010-03-03 24 -177 val_177 2010-03-03 24 -452 val_452 2010-03-03 24 -242 val_242 2010-03-03 24 -401 val_401 2010-03-03 24 -331 val_331 2010-03-03 24 +278 val_278 2010-03-03 24 +176 val_176 2010-03-03 24 +289 val_289 2010-03-03 24 +221 val_221 2010-03-03 24 +65 val_65 2010-03-03 24 +318 val_318 2010-03-03 24 +332 val_332 2010-03-03 24 +311 val_311 2010-03-03 24 +275 val_275 2010-03-03 24 +137 val_137 2010-03-03 24 +241 val_241 2010-03-03 24 +83 val_83 2010-03-03 24 +333 val_333 2010-03-03 24 +180 val_180 2010-03-03 24 +284 val_284 2010-03-03 24 +12 val_12 2010-03-03 24 +230 val_230 2010-03-03 24 +181 val_181 2010-03-03 24 +67 val_67 2010-03-03 24 +260 val_260 2010-03-03 24 +404 val_404 2010-03-03 24 +384 val_384 2010-03-03 24 +489 val_489 2010-03-03 24 +353 val_353 2010-03-03 24 +373 val_373 2010-03-03 24 272 val_272 2010-03-03 24 -392 val_392 2010-03-03 24 -369 val_369 2010-03-03 24 -242 val_242 2010-03-03 24 -327 val_327 2010-03-03 24 -389 val_389 2010-03-03 24 -309 val_309 2010-03-03 24 -224 val_224 2010-03-03 24 -306 val_306 2010-03-03 24 -273 val_273 2010-03-03 24 -277 val_277 2010-03-03 24 -10 val_10 2010-03-03 24 -435 val_435 2010-03-03 24 -19 val_19 2010-03-03 24 -307 val_307 2010-03-03 24 -90 val_90 2010-03-03 24 -72 val_72 2010-03-03 24 -118 val_118 2010-03-03 24 -15 val_15 2010-03-03 24 -419 val_419 2010-03-03 24 -238 val_238 2010-03-03 24 138 val_138 2010-03-03 24 -282 val_282 2010-03-03 24 -395 val_395 2010-03-03 24 -134 val_134 2010-03-03 24 -118 val_118 2010-03-03 24 +217 val_217 2010-03-03 24 +84 val_84 2010-03-03 24 +348 val_348 2010-03-03 24 +466 val_466 2010-03-03 24 +58 val_58 2010-03-03 24 +8 val_8 2010-03-03 24 +411 val_411 2010-03-03 24 +230 val_230 2010-03-03 24 +208 val_208 2010-03-03 24 +348 val_348 2010-03-03 24 +24 val_24 2010-03-03 24 +463 val_463 2010-03-03 24 +431 val_431 2010-03-03 24 179 val_179 2010-03-03 24 -364 val_364 2010-03-03 24 -87 val_87 2010-03-03 24 -490 val_490 2010-03-03 24 +172 val_172 2010-03-03 24 +42 val_42 2010-03-03 24 +129 val_129 2010-03-03 24 +158 val_158 2010-03-03 24 +119 val_119 2010-03-03 24 +496 val_496 2010-03-03 24 0 val_0 2010-03-03 24 -77 val_77 2010-03-03 24 -85 val_85 2010-03-03 24 -413 val_413 2010-03-03 24 -169 val_169 2010-03-03 24 -316 val_316 2010-03-03 24 -369 val_369 2010-03-03 24 -137 val_137 2010-03-03 24 -470 val_470 2010-03-03 24 -409 val_409 2010-03-03 24 -318 val_318 2010-03-03 24 -318 val_318 2010-03-03 24 -187 val_187 2010-03-03 24 +322 val_322 2010-03-03 24 197 val_197 2010-03-03 24 -282 val_282 2010-03-03 24 -98 val_98 2010-03-03 24 -457 val_457 2010-03-03 24 -481 val_481 2010-03-03 24 -288 val_288 2010-03-03 24 -196 val_196 2010-03-03 24 -95 val_95 2010-03-03 24 -308 val_308 2010-03-03 24 468 val_468 2010-03-03 24 -469 val_469 2010-03-03 24 -156 val_156 2010-03-03 24 -436 val_436 2010-03-03 24 -43 val_43 2010-03-03 24 -404 val_404 2010-03-03 24 -51 val_51 2010-03-03 24 -131 val_131 2010-03-03 24 -120 val_120 2010-03-03 24 -205 val_205 2010-03-03 24 -230 val_230 2010-03-03 24 -327 val_327 2010-03-03 24 -165 val_165 2010-03-03 24 -26 val_26 2010-03-03 24 -96 val_96 2010-03-03 24 -418 val_418 2010-03-03 24 -191 val_191 2010-03-03 24 -199 val_199 2010-03-03 24 -298 val_298 2010-03-03 24 -100 val_100 2010-03-03 24 -454 val_454 2010-03-03 24 393 val_393 2010-03-03 24 +454 val_454 2010-03-03 24 +100 val_100 2010-03-03 24 +298 val_298 2010-03-03 24 +199 val_199 2010-03-03 24 +191 val_191 2010-03-03 24 +418 val_418 2010-03-03 24 +96 val_96 2010-03-03 24 +26 val_26 2010-03-03 24 +165 val_165 2010-03-03 24 +327 val_327 2010-03-03 24 +230 val_230 2010-03-03 24 +205 val_205 2010-03-03 24 +120 val_120 2010-03-03 24 +131 val_131 2010-03-03 24 +51 val_51 2010-03-03 24 +404 val_404 2010-03-03 24 +43 val_43 2010-03-03 24 +436 val_436 2010-03-03 24 +156 val_156 2010-03-03 24 +469 val_469 2010-03-03 24 468 val_468 2010-03-03 24 +308 val_308 2010-03-03 24 +95 val_95 2010-03-03 24 +196 val_196 2010-03-03 24 +288 val_288 2010-03-03 24 +481 val_481 2010-03-03 24 +457 val_457 2010-03-03 24 +98 val_98 2010-03-03 24 +282 val_282 2010-03-03 24 197 val_197 2010-03-03 24 +187 val_187 2010-03-03 24 +318 val_318 2010-03-03 24 +318 val_318 2010-03-03 24 +409 val_409 2010-03-03 24 +470 val_470 2010-03-03 24 +137 val_137 2010-03-03 24 +369 val_369 2010-03-03 24 +316 val_316 2010-03-03 24 +169 val_169 2010-03-03 24 +413 val_413 2010-03-03 24 +85 val_85 2010-03-03 24 +77 val_77 2010-03-03 24 +0 val_0 2010-03-03 24 +490 val_490 2010-03-03 24 +87 val_87 2010-03-03 24 +364 val_364 2010-03-03 24 +179 val_179 2010-03-03 24 +118 val_118 2010-03-03 24 +134 val_134 2010-03-03 24 +395 val_395 2010-03-03 24 +282 val_282 2010-03-03 24 +138 val_138 2010-03-03 24 +238 val_238 2010-03-03 24 +419 val_419 2010-03-03 24 +15 val_15 2010-03-03 24 +118 val_118 2010-03-03 24 +72 val_72 2010-03-03 24 +90 val_90 2010-03-03 24 +307 val_307 2010-03-03 24 +19 val_19 2010-03-03 24 +435 val_435 2010-03-03 24 +10 val_10 2010-03-03 24 +277 val_277 2010-03-03 24 +273 val_273 2010-03-03 24 +306 val_306 2010-03-03 24 +224 val_224 2010-03-03 24 +309 val_309 2010-03-03 24 +389 val_389 2010-03-03 24 +327 val_327 2010-03-03 24 +242 val_242 2010-03-03 24 +369 val_369 2010-03-03 24 +392 val_392 2010-03-03 24 +272 val_272 2010-03-03 24 +331 val_331 2010-03-03 24 +401 val_401 2010-03-03 24 +242 val_242 2010-03-03 24 +452 val_452 2010-03-03 24 +177 val_177 2010-03-03 24 +226 val_226 2010-03-03 24 +5 val_5 2010-03-03 24 +497 val_497 2010-03-03 24 +402 val_402 2010-03-03 24 +396 val_396 2010-03-03 24 +317 val_317 2010-03-03 24 +395 val_395 2010-03-03 24 +58 val_58 2010-03-03 24 +35 val_35 2010-03-03 24 +336 val_336 2010-03-03 24 +95 val_95 2010-03-03 24 +11 val_11 2010-03-03 24 +168 val_168 2010-03-03 24 +34 val_34 2010-03-03 24 +229 val_229 2010-03-03 24 +233 val_233 2010-03-03 24 +143 val_143 2010-03-03 24 +472 val_472 2010-03-03 24 322 val_322 2010-03-03 24 -0 val_0 2010-03-03 24 -496 val_496 2010-03-03 24 -119 val_119 2010-03-03 24 -158 val_158 2010-03-03 24 -129 val_129 2010-03-03 24 +498 val_498 2010-03-03 24 +160 val_160 2010-03-03 24 +195 val_195 2010-03-03 24 42 val_42 2010-03-03 24 -172 val_172 2010-03-03 24 -179 val_179 2010-03-03 24 -431 val_431 2010-03-03 24 -463 val_463 2010-03-03 24 -24 val_24 2010-03-03 24 -348 val_348 2010-03-03 24 -208 val_208 2010-03-03 24 -230 val_230 2010-03-03 24 -411 val_411 2010-03-03 24 -8 val_8 2010-03-03 24 -58 val_58 2010-03-03 24 -466 val_466 2010-03-03 24 -348 val_348 2010-03-03 24 -84 val_84 2010-03-03 24 -217 val_217 2010-03-03 24 -138 val_138 2010-03-03 24 -272 val_272 2010-03-03 24 -373 val_373 2010-03-03 24 -353 val_353 2010-03-03 24 +321 val_321 2010-03-03 24 +430 val_430 2010-03-03 24 +119 val_119 2010-03-03 24 489 val_489 2010-03-03 24 -384 val_384 2010-03-03 24 -404 val_404 2010-03-03 24 -260 val_260 2010-03-03 24 -67 val_67 2010-03-03 24 -181 val_181 2010-03-03 24 +458 val_458 2010-03-03 24 +78 val_78 2010-03-03 24 +76 val_76 2010-03-03 24 +41 val_41 2010-03-03 24 +223 val_223 2010-03-03 24 +492 val_492 2010-03-03 24 +149 val_149 2010-03-03 24 +449 val_449 2010-03-03 24 +218 val_218 2010-03-03 24 +228 val_228 2010-03-03 24 +138 val_138 2010-03-03 24 +453 val_453 2010-03-03 24 +30 val_30 2010-03-03 24 +209 val_209 2010-03-03 24 +64 val_64 2010-03-03 24 +468 val_468 2010-03-03 24 +76 val_76 2010-03-03 24 +74 val_74 2010-03-03 24 +342 val_342 2010-03-03 24 +69 val_69 2010-03-03 24 230 val_230 2010-03-03 24 -12 val_12 2010-03-03 24 -284 val_284 2010-03-03 24 -180 val_180 2010-03-03 24 -333 val_333 2010-03-03 24 -83 val_83 2010-03-03 24 -241 val_241 2010-03-03 24 -137 val_137 2010-03-03 24 -275 val_275 2010-03-03 24 -311 val_311 2010-03-03 24 -332 val_332 2010-03-03 24 -318 val_318 2010-03-03 24 -65 val_65 2010-03-03 24 -221 val_221 2010-03-03 24 -289 val_289 2010-03-03 24 -176 val_176 2010-03-03 24 -278 val_278 2010-03-03 24 -430 val_430 2010-03-03 24 +33 val_33 2010-03-03 24 +368 val_368 2010-03-03 24 +103 val_103 2010-03-03 24 +296 val_296 2010-03-03 24 +113 val_113 2010-03-03 24 216 val_216 2010-03-03 24 -213 val_213 2010-03-03 24 +367 val_367 2010-03-03 24 +344 val_344 2010-03-03 24 +167 val_167 2010-03-03 24 +274 val_274 2010-03-03 24 +219 val_219 2010-03-03 24 239 val_239 2010-03-03 24 -103 val_103 2010-03-03 24 -138 val_138 2010-03-03 24 -51 val_51 2010-03-03 24 -459 val_459 2010-03-03 24 -54 val_54 2010-03-03 24 -176 val_176 2010-03-03 24 -187 val_187 2010-03-03 24 -286 val_286 2010-03-03 24 -192 val_192 2010-03-03 24 -469 val_469 2010-03-03 24 -437 val_437 2010-03-03 24 -386 val_386 2010-03-03 24 -125 val_125 2010-03-03 24 -498 val_498 2010-03-03 24 -382 val_382 2010-03-03 24 -169 val_169 2010-03-03 24 -399 val_399 2010-03-03 24 -356 val_356 2010-03-03 24 -208 val_208 2010-03-03 24 -277 val_277 2010-03-03 24 -427 val_427 2010-03-03 24 -35 val_35 2010-03-03 24 -280 val_280 2010-03-03 24 -4 val_4 2010-03-03 24 -72 val_72 2010-03-03 24 -47 val_47 2010-03-03 24 -111 val_111 2010-03-03 24 -92 val_92 2010-03-03 24 -221 val_221 2010-03-03 24 -378 val_378 2010-03-03 24 -157 val_157 2010-03-03 24 -489 val_489 2010-03-03 24 -20 val_20 2010-03-03 24 -170 val_170 2010-03-03 24 -129 val_129 2010-03-03 24 -345 val_345 2010-03-03 24 +485 val_485 2010-03-03 24 +116 val_116 2010-03-03 24 +223 val_223 2010-03-03 24 +256 val_256 2010-03-03 24 +263 val_263 2010-03-03 24 +70 val_70 2010-03-03 24 +487 val_487 2010-03-03 24 +480 val_480 2010-03-03 24 +401 val_401 2010-03-03 24 +288 val_288 2010-03-03 24 +191 val_191 2010-03-03 24 +5 val_5 2010-03-03 24 +244 val_244 2010-03-03 24 438 val_438 2010-03-03 24 -149 val_149 2010-03-03 24 -205 val_205 2010-03-03 24 -302 val_302 2010-03-03 24 -57 val_57 2010-03-03 24 -316 val_316 2010-03-03 24 -311 val_311 2010-03-03 24 128 val_128 2010-03-03 24 -455 val_455 2010-03-03 24 -0 val_0 2010-03-03 24 -339 val_339 2010-03-03 24 +467 val_467 2010-03-03 24 +432 val_432 2010-03-03 24 +202 val_202 2010-03-03 24 +316 val_316 2010-03-03 24 +229 val_229 2010-03-03 24 +469 val_469 2010-03-03 24 +463 val_463 2010-03-03 24 +280 val_280 2010-03-03 24 +2 val_2 2010-03-03 24 +35 val_35 2010-03-03 24 +283 val_283 2010-03-03 24 +331 val_331 2010-03-03 24 +235 val_235 2010-03-03 24 +80 val_80 2010-03-03 24 +44 val_44 2010-03-03 24 +193 val_193 2010-03-03 24 +321 val_321 2010-03-03 24 +335 val_335 2010-03-03 24 +104 val_104 2010-03-03 24 +466 val_466 2010-03-03 24 +366 val_366 2010-03-03 24 +175 val_175 2010-03-03 24 +403 val_403 2010-03-03 24 +483 val_483 2010-03-03 24 +53 val_53 2010-03-03 24 +105 val_105 2010-03-03 24 +257 val_257 2010-03-03 24 +406 val_406 2010-03-03 24 +409 val_409 2010-03-03 24 +190 val_190 2010-03-03 24 +406 val_406 2010-03-03 24 +401 val_401 2010-03-03 24 +114 val_114 2010-03-03 24 +258 val_258 2010-03-03 24 +90 val_90 2010-03-03 24 203 val_203 2010-03-03 24 -155 val_155 2010-03-03 24 -113 val_113 2010-03-03 24 -17 val_17 2010-03-03 24 -475 val_475 2010-03-03 24 -195 val_195 2010-03-03 24 -167 val_167 2010-03-03 24 -325 val_325 2010-03-03 24 -367 val_367 2010-03-03 24 -342 val_342 2010-03-03 24 -439 val_439 2010-03-03 24 -266 val_266 2010-03-03 24 -365 val_365 2010-03-03 24 -309 val_309 2010-03-03 24 -397 val_397 2010-03-03 24 -377 val_377 2010-03-03 24 -162 val_162 2010-03-03 24 -489 val_489 2010-03-03 24 -417 val_417 2010-03-03 24 -247 val_247 2010-03-03 24 +262 val_262 2010-03-03 24 +348 val_348 2010-03-03 24 +424 val_424 2010-03-03 24 +12 val_12 2010-03-03 24 396 val_396 2010-03-03 24 -399 val_399 2010-03-03 24 -174 val_174 2010-03-03 24 -208 val_208 2010-03-03 24 -466 val_466 2010-03-03 24 -199 val_199 2010-03-03 24 -207 val_207 2010-03-03 24 -494 val_494 2010-03-03 24 -413 val_413 2010-03-03 24 -174 val_174 2010-03-03 24 -482 val_482 2010-03-03 24 +201 val_201 2010-03-03 24 +217 val_217 2010-03-03 24 +164 val_164 2010-03-03 24 +431 val_431 2010-03-03 24 +454 val_454 2010-03-03 24 +478 val_478 2010-03-03 24 +298 val_298 2010-03-03 24 +125 val_125 2010-03-03 24 +431 val_431 2010-03-03 24 +164 val_164 2010-03-03 24 +424 val_424 2010-03-03 24 +187 val_187 2010-03-03 24 +382 val_382 2010-03-03 24 +5 val_5 2010-03-03 24 +70 val_70 2010-03-03 24 +397 val_397 2010-03-03 24 +480 val_480 2010-03-03 24 +291 val_291 2010-03-03 24 +24 val_24 2010-03-03 24 +351 val_351 2010-03-03 24 +255 val_255 2010-03-03 24 +104 val_104 2010-03-03 24 +70 val_70 2010-03-03 24 +163 val_163 2010-03-03 24 +438 val_438 2010-03-03 24 +119 val_119 2010-03-03 24 +414 val_414 2010-03-03 24 +200 val_200 2010-03-03 24 +491 val_491 2010-03-03 24 237 val_237 2010-03-03 24 -394 val_394 2010-03-03 24 -459 val_459 2010-03-03 24 -446 val_446 2010-03-03 24 -338 val_338 2010-03-03 24 -193 val_193 2010-03-03 24 -153 val_153 2010-03-03 24 -287 val_287 2010-03-03 24 -219 val_219 2010-03-03 24 -292 val_292 2010-03-03 24 -252 val_252 2010-03-03 24 -430 val_430 2010-03-03 24 +439 val_439 2010-03-03 24 +360 val_360 2010-03-03 24 +248 val_248 2010-03-03 24 +479 val_479 2010-03-03 24 +305 val_305 2010-03-03 24 417 val_417 2010-03-03 24 -166 val_166 2010-03-03 24 -403 val_403 2010-03-03 24 -82 val_82 2010-03-03 24 -15 val_15 2010-03-03 24 -209 val_209 2010-03-03 24 -277 val_277 2010-03-03 24 -281 val_281 2010-03-03 24 -327 val_327 2010-03-03 24 -37 val_37 2010-03-03 24 -495 val_495 2010-03-03 24 -145 val_145 2010-03-03 24 -469 val_469 2010-03-03 24 -152 val_152 2010-03-03 24 -374 val_374 2010-03-03 24 +199 val_199 2010-03-03 24 +444 val_444 2010-03-03 24 +120 val_120 2010-03-03 24 429 val_429 2010-03-03 24 +169 val_169 2010-03-03 24 +443 val_443 2010-03-03 24 +323 val_323 2010-03-03 24 +325 val_325 2010-03-03 24 +277 val_277 2010-03-03 24 +230 val_230 2010-03-03 24 +478 val_478 2010-03-03 24 +178 val_178 2010-03-03 24 +468 val_468 2010-03-03 24 +310 val_310 2010-03-03 24 +317 val_317 2010-03-03 24 +333 val_333 2010-03-03 24 +493 val_493 2010-03-03 24 +460 val_460 2010-03-03 24 +207 val_207 2010-03-03 24 +249 val_249 2010-03-03 24 +265 val_265 2010-03-03 24 +480 val_480 2010-03-03 24 +83 val_83 2010-03-03 24 +136 val_136 2010-03-03 24 +353 val_353 2010-03-03 24 +172 val_172 2010-03-03 24 +214 val_214 2010-03-03 24 +462 val_462 2010-03-03 24 +233 val_233 2010-03-03 24 406 val_406 2010-03-03 24 -146 val_146 2010-03-03 24 -213 val_213 2010-03-03 24 -128 val_128 2010-03-03 24 -66 val_66 2010-03-03 24 -369 val_369 2010-03-03 24 -224 val_224 2010-03-03 24 -273 val_273 2010-03-03 24 -150 val_150 2010-03-03 24 +133 val_133 2010-03-03 24 +175 val_175 2010-03-03 24 +189 val_189 2010-03-03 24 +454 val_454 2010-03-03 24 +375 val_375 2010-03-03 24 401 val_401 2010-03-03 24 -193 val_193 2010-03-03 24 -265 val_265 2010-03-03 24 -484 val_484 2010-03-03 24 -98 val_98 2010-03-03 24 -278 val_278 2010-03-03 24 -255 val_255 2010-03-03 24 -409 val_409 2010-03-03 24 -165 val_165 2010-03-03 24 -27 val_27 2010-03-03 24 -311 val_311 2010-03-03 24 -86 val_86 2010-03-03 24 -238 val_238 2010-03-03 24 -200 val_200 2010-03-03 24 -400 val_400 2010-03-03 24 -403 val_403 2010-03-03 24 -169 val_169 2010-03-03 24 -90 val_90 2010-03-03 24 -126 val_126 2010-03-03 24 -222 val_222 2010-03-03 24 -477 val_477 2010-03-03 24 -414 val_414 2010-03-03 24 -194 val_194 2010-03-03 24 -307 val_307 2010-03-03 24 +421 val_421 2010-03-03 24 +407 val_407 2010-03-03 24 +384 val_384 2010-03-03 24 +256 val_256 2010-03-03 24 +26 val_26 2010-03-03 24 +134 val_134 2010-03-03 24 +67 val_67 2010-03-03 24 +384 val_384 2010-03-03 24 +379 val_379 2010-03-03 24 +18 val_18 2010-03-03 24 +462 val_462 2010-03-03 24 +492 val_492 2010-03-03 24 +100 val_100 2010-03-03 24 +298 val_298 2010-03-03 24 +9 val_9 2010-03-03 24 +341 val_341 2010-03-03 24 +498 val_498 2010-03-03 24 +146 val_146 2010-03-03 24 +458 val_458 2010-03-03 24 +362 val_362 2010-03-03 24 +186 val_186 2010-03-03 24 +285 val_285 2010-03-03 24 348 val_348 2010-03-03 24 -152 val_152 2010-03-03 24 -448 val_448 2010-03-03 24 -37 val_37 2010-03-03 24 -28 val_28 2010-03-03 24 -84 val_84 2010-03-03 24 -315 val_315 2010-03-03 24 +167 val_167 2010-03-03 24 +18 val_18 2010-03-03 24 +273 val_273 2010-03-03 24 +183 val_183 2010-03-03 24 +281 val_281 2010-03-03 24 +344 val_344 2010-03-03 24 +97 val_97 2010-03-03 24 469 val_469 2010-03-03 24 +315 val_315 2010-03-03 24 +84 val_84 2010-03-03 24 +28 val_28 2010-03-03 24 +37 val_37 2010-03-03 24 +448 val_448 2010-03-03 24 +152 val_152 2010-03-03 24 +348 val_348 2010-03-03 24 +307 val_307 2010-03-03 24 +194 val_194 2010-03-03 24 +414 val_414 2010-03-03 24 +477 val_477 2010-03-03 24 +222 val_222 2010-03-03 24 +126 val_126 2010-03-03 24 +90 val_90 2010-03-03 24 +169 val_169 2010-03-03 24 +403 val_403 2010-03-03 24 +400 val_400 2010-03-03 24 +200 val_200 2010-03-03 24 97 val_97 2010-03-03 24 -344 val_344 2010-03-03 24 -281 val_281 2010-03-03 24 -183 val_183 2010-03-03 24 +238 val_238 2010-03-03 24 +86 val_86 2010-03-03 24 +311 val_311 2010-03-03 24 +27 val_27 2010-03-03 24 +165 val_165 2010-03-03 24 +409 val_409 2010-03-03 24 +255 val_255 2010-03-03 24 +278 val_278 2010-03-03 24 +98 val_98 2010-03-03 24 +484 val_484 2010-03-03 24 +265 val_265 2010-03-03 24 +193 val_193 2010-03-03 24 +401 val_401 2010-03-03 24 +150 val_150 2010-03-03 24 273 val_273 2010-03-03 24 -18 val_18 2010-03-03 24 -167 val_167 2010-03-03 24 -348 val_348 2010-03-03 24 -285 val_285 2010-03-03 24 -186 val_186 2010-03-03 24 -362 val_362 2010-03-03 24 -458 val_458 2010-03-03 24 +224 val_224 2010-03-03 24 +369 val_369 2010-03-03 24 +66 val_66 2010-03-03 24 +128 val_128 2010-03-03 24 +213 val_213 2010-03-03 24 146 val_146 2010-03-03 24 -498 val_498 2010-03-03 24 -341 val_341 2010-03-03 24 -9 val_9 2010-03-03 24 -298 val_298 2010-03-03 24 -100 val_100 2010-03-03 24 -492 val_492 2010-03-03 24 -462 val_462 2010-03-03 24 -18 val_18 2010-03-03 24 -379 val_379 2010-03-03 24 -384 val_384 2010-03-03 24 -67 val_67 2010-03-03 24 -134 val_134 2010-03-03 24 -26 val_26 2010-03-03 24 -256 val_256 2010-03-03 24 -384 val_384 2010-03-03 24 -407 val_407 2010-03-03 24 -421 val_421 2010-03-03 24 -401 val_401 2010-03-03 24 -375 val_375 2010-03-03 24 -454 val_454 2010-03-03 24 -189 val_189 2010-03-03 24 -175 val_175 2010-03-03 24 -133 val_133 2010-03-03 24 406 val_406 2010-03-03 24 -233 val_233 2010-03-03 24 -462 val_462 2010-03-03 24 -214 val_214 2010-03-03 24 -172 val_172 2010-03-03 24 -353 val_353 2010-03-03 24 -136 val_136 2010-03-03 24 -83 val_83 2010-03-03 24 -480 val_480 2010-03-03 24 -265 val_265 2010-03-03 24 -249 val_249 2010-03-03 24 -207 val_207 2010-03-03 24 -460 val_460 2010-03-03 24 -493 val_493 2010-03-03 24 -333 val_333 2010-03-03 24 -317 val_317 2010-03-03 24 -310 val_310 2010-03-03 24 -468 val_468 2010-03-03 24 -178 val_178 2010-03-03 24 -478 val_478 2010-03-03 24 -230 val_230 2010-03-03 24 -277 val_277 2010-03-03 24 -325 val_325 2010-03-03 24 -323 val_323 2010-03-03 24 -443 val_443 2010-03-03 24 -169 val_169 2010-03-03 24 429 val_429 2010-03-03 24 -120 val_120 2010-03-03 24 -444 val_444 2010-03-03 24 +374 val_374 2010-03-03 24 +152 val_152 2010-03-03 24 +469 val_469 2010-03-03 24 +145 val_145 2010-03-03 24 +495 val_495 2010-03-03 24 +37 val_37 2010-03-03 24 +327 val_327 2010-03-03 24 +281 val_281 2010-03-03 24 +277 val_277 2010-03-03 24 +209 val_209 2010-03-03 24 +15 val_15 2010-03-03 24 +82 val_82 2010-03-03 24 +403 val_403 2010-03-03 24 +166 val_166 2010-03-03 24 +417 val_417 2010-03-03 24 +430 val_430 2010-03-03 24 +252 val_252 2010-03-03 24 +292 val_292 2010-03-03 24 +219 val_219 2010-03-03 24 +287 val_287 2010-03-03 24 +153 val_153 2010-03-03 24 +193 val_193 2010-03-03 24 +338 val_338 2010-03-03 24 +446 val_446 2010-03-03 24 +459 val_459 2010-03-03 24 +394 val_394 2010-03-03 24 +237 val_237 2010-03-03 24 +482 val_482 2010-03-03 24 +174 val_174 2010-03-03 24 +413 val_413 2010-03-03 24 +494 val_494 2010-03-03 24 +207 val_207 2010-03-03 24 199 val_199 2010-03-03 24 +466 val_466 2010-03-03 24 +208 val_208 2010-03-03 24 +174 val_174 2010-03-03 24 +399 val_399 2010-03-03 24 +396 val_396 2010-03-03 24 +247 val_247 2010-03-03 24 417 val_417 2010-03-03 24 -305 val_305 2010-03-03 24 -479 val_479 2010-03-03 24 -248 val_248 2010-03-03 24 -360 val_360 2010-03-03 24 +489 val_489 2010-03-03 24 +162 val_162 2010-03-03 24 +377 val_377 2010-03-03 24 +397 val_397 2010-03-03 24 +309 val_309 2010-03-03 24 +365 val_365 2010-03-03 24 +266 val_266 2010-03-03 24 439 val_439 2010-03-03 24 -237 val_237 2010-03-03 24 -491 val_491 2010-03-03 24 -200 val_200 2010-03-03 24 -414 val_414 2010-03-03 24 -119 val_119 2010-03-03 24 +342 val_342 2010-03-03 24 +367 val_367 2010-03-03 24 +325 val_325 2010-03-03 24 +167 val_167 2010-03-03 24 +195 val_195 2010-03-03 24 +475 val_475 2010-03-03 24 +17 val_17 2010-03-03 24 +113 val_113 2010-03-03 24 +155 val_155 2010-03-03 24 +203 val_203 2010-03-03 24 +339 val_339 2010-03-03 24 +0 val_0 2010-03-03 24 +455 val_455 2010-03-03 24 +128 val_128 2010-03-03 24 +311 val_311 2010-03-03 24 +316 val_316 2010-03-03 24 +57 val_57 2010-03-03 24 +302 val_302 2010-03-03 24 +205 val_205 2010-03-03 24 +149 val_149 2010-03-03 24 438 val_438 2010-03-03 24 -163 val_163 2010-03-03 24 -70 val_70 2010-03-03 24 -104 val_104 2010-03-03 24 -255 val_255 2010-03-03 24 -351 val_351 2010-03-03 24 -24 val_24 2010-03-03 24 -291 val_291 2010-03-03 24 -480 val_480 2010-03-03 24 -397 val_397 2010-03-03 24 -70 val_70 2010-03-03 24 -5 val_5 2010-03-03 24 +345 val_345 2010-03-03 24 +129 val_129 2010-03-03 24 +170 val_170 2010-03-03 24 +20 val_20 2010-03-03 24 +489 val_489 2010-03-03 24 +157 val_157 2010-03-03 24 +378 val_378 2010-03-03 24 +221 val_221 2010-03-03 24 +92 val_92 2010-03-03 24 +111 val_111 2010-03-03 24 +47 val_47 2010-03-03 24 +72 val_72 2010-03-03 24 +4 val_4 2010-03-03 24 +280 val_280 2010-03-03 24 +35 val_35 2010-03-03 24 +427 val_427 2010-03-03 24 +277 val_277 2010-03-03 24 +208 val_208 2010-03-03 24 +356 val_356 2010-03-03 24 +399 val_399 2010-03-03 24 +169 val_169 2010-03-03 24 382 val_382 2010-03-03 24 -187 val_187 2010-03-03 24 -424 val_424 2010-03-03 24 -164 val_164 2010-03-03 24 -431 val_431 2010-03-03 24 +498 val_498 2010-03-03 24 125 val_125 2010-03-03 24 -298 val_298 2010-03-03 24 -478 val_478 2010-03-03 24 -454 val_454 2010-03-03 24 -431 val_431 2010-03-03 24 -164 val_164 2010-03-03 24 -217 val_217 2010-03-03 24 -201 val_201 2010-03-03 24 -396 val_396 2010-03-03 24 -12 val_12 2010-03-03 24 -424 val_424 2010-03-03 24 -348 val_348 2010-03-03 24 -262 val_262 2010-03-03 24 -203 val_203 2010-03-03 24 -90 val_90 2010-03-03 24 -258 val_258 2010-03-03 24 -114 val_114 2010-03-03 24 -401 val_401 2010-03-03 24 -406 val_406 2010-03-03 24 -190 val_190 2010-03-03 24 -409 val_409 2010-03-03 24 -406 val_406 2010-03-03 24 -257 val_257 2010-03-03 24 -105 val_105 2010-03-03 24 -53 val_53 2010-03-03 24 -483 val_483 2010-03-03 24 -403 val_403 2010-03-03 24 -175 val_175 2010-03-03 24 -366 val_366 2010-03-03 24 -466 val_466 2010-03-03 24 -104 val_104 2010-03-03 24 -335 val_335 2010-03-03 24 -321 val_321 2010-03-03 24 -193 val_193 2010-03-03 24 -44 val_44 2010-03-03 24 -80 val_80 2010-03-03 24 -235 val_235 2010-03-03 24 -331 val_331 2010-03-03 24 -283 val_283 2010-03-03 24 -35 val_35 2010-03-03 24 -2 val_2 2010-03-03 24 -280 val_280 2010-03-03 24 -463 val_463 2010-03-03 24 +386 val_386 2010-03-03 24 +437 val_437 2010-03-03 24 469 val_469 2010-03-03 24 -229 val_229 2010-03-03 24 -316 val_316 2010-03-03 24 -202 val_202 2010-03-03 24 -432 val_432 2010-03-03 24 -467 val_467 2010-03-03 24 -128 val_128 2010-03-03 24 -438 val_438 2010-03-03 24 -244 val_244 2010-03-03 24 -5 val_5 2010-03-03 24 -191 val_191 2010-03-03 24 -288 val_288 2010-03-03 24 -401 val_401 2010-03-03 24 -480 val_480 2010-03-03 24 -487 val_487 2010-03-03 24 -70 val_70 2010-03-03 24 -263 val_263 2010-03-03 24 -256 val_256 2010-03-03 24 -223 val_223 2010-03-03 24 -116 val_116 2010-03-03 24 -485 val_485 2010-03-03 24 +192 val_192 2010-03-03 24 +286 val_286 2010-03-03 24 +187 val_187 2010-03-03 24 +176 val_176 2010-03-03 24 +54 val_54 2010-03-03 24 +459 val_459 2010-03-03 24 +51 val_51 2010-03-03 24 +138 val_138 2010-03-03 24 +103 val_103 2010-03-03 24 239 val_239 2010-03-03 24 -219 val_219 2010-03-03 24 -274 val_274 2010-03-03 24 -167 val_167 2010-03-03 24 -344 val_344 2010-03-03 24 -367 val_367 2010-03-03 24 +213 val_213 2010-03-03 24 216 val_216 2010-03-03 24 -113 val_113 2010-03-03 24 -296 val_296 2010-03-03 24 -103 val_103 2010-03-03 24 -368 val_368 2010-03-03 24 -33 val_33 2010-03-03 24 +430 val_430 2010-03-03 24 +278 val_278 2010-03-03 24 +176 val_176 2010-03-03 24 +289 val_289 2010-03-03 24 +221 val_221 2010-03-03 24 +65 val_65 2010-03-03 24 +318 val_318 2010-03-03 24 +332 val_332 2010-03-03 24 +311 val_311 2010-03-03 24 +275 val_275 2010-03-03 24 +137 val_137 2010-03-03 24 +241 val_241 2010-03-03 24 +83 val_83 2010-03-03 24 +333 val_333 2010-03-03 24 +180 val_180 2010-03-03 24 +284 val_284 2010-03-03 24 +12 val_12 2010-03-03 24 230 val_230 2010-03-03 24 -69 val_69 2010-03-03 24 -342 val_342 2010-03-03 24 -74 val_74 2010-03-03 24 -76 val_76 2010-03-03 24 -468 val_468 2010-03-03 24 -64 val_64 2010-03-03 24 -209 val_209 2010-03-03 24 -30 val_30 2010-03-03 24 -453 val_453 2010-03-03 24 -138 val_138 2010-03-03 24 -228 val_228 2010-03-03 24 -218 val_218 2010-03-03 24 -449 val_449 2010-03-03 24 -149 val_149 2010-03-03 24 -492 val_492 2010-03-03 24 -223 val_223 2010-03-03 24 -41 val_41 2010-03-03 24 -76 val_76 2010-03-03 24 -78 val_78 2010-03-03 24 -458 val_458 2010-03-03 24 +181 val_181 2010-03-03 24 +67 val_67 2010-03-03 24 +260 val_260 2010-03-03 24 +404 val_404 2010-03-03 24 +384 val_384 2010-03-03 24 489 val_489 2010-03-03 24 -119 val_119 2010-03-03 24 -430 val_430 2010-03-03 24 -321 val_321 2010-03-03 24 -42 val_42 2010-03-03 24 -195 val_195 2010-03-03 24 -160 val_160 2010-03-03 24 -498 val_498 2010-03-03 24 -322 val_322 2010-03-03 24 -472 val_472 2010-03-03 24 -143 val_143 2010-03-03 24 -233 val_233 2010-03-03 24 -229 val_229 2010-03-03 24 -34 val_34 2010-03-03 24 -168 val_168 2010-03-03 24 -11 val_11 2010-03-03 24 -95 val_95 2010-03-03 24 -336 val_336 2010-03-03 24 -35 val_35 2010-03-03 24 -58 val_58 2010-03-03 24 -395 val_395 2010-03-03 24 -317 val_317 2010-03-03 24 -396 val_396 2010-03-03 24 -402 val_402 2010-03-03 24 -497 val_497 2010-03-03 24 -5 val_5 2010-03-03 24 -226 val_226 2010-03-03 24 -177 val_177 2010-03-03 24 -452 val_452 2010-03-03 24 -242 val_242 2010-03-03 24 -401 val_401 2010-03-03 24 -331 val_331 2010-03-03 24 +353 val_353 2010-03-03 24 +373 val_373 2010-03-03 24 272 val_272 2010-03-03 24 -392 val_392 2010-03-03 24 -369 val_369 2010-03-03 24 -242 val_242 2010-03-03 24 -327 val_327 2010-03-03 24 -389 val_389 2010-03-03 24 -309 val_309 2010-03-03 24 -224 val_224 2010-03-03 24 -306 val_306 2010-03-03 24 -273 val_273 2010-03-03 24 -277 val_277 2010-03-03 24 -10 val_10 2010-03-03 24 -435 val_435 2010-03-03 24 -19 val_19 2010-03-03 24 -307 val_307 2010-03-03 24 -90 val_90 2010-03-03 24 -72 val_72 2010-03-03 24 -118 val_118 2010-03-03 24 -15 val_15 2010-03-03 24 -419 val_419 2010-03-03 24 -238 val_238 2010-03-03 24 138 val_138 2010-03-03 24 -282 val_282 2010-03-03 24 -395 val_395 2010-03-03 24 -134 val_134 2010-03-03 24 -118 val_118 2010-03-03 24 +217 val_217 2010-03-03 24 +84 val_84 2010-03-03 24 +348 val_348 2010-03-03 24 +466 val_466 2010-03-03 24 +58 val_58 2010-03-03 24 +8 val_8 2010-03-03 24 +411 val_411 2010-03-03 24 +230 val_230 2010-03-03 24 +208 val_208 2010-03-03 24 +348 val_348 2010-03-03 24 +24 val_24 2010-03-03 24 +463 val_463 2010-03-03 24 +431 val_431 2010-03-03 24 179 val_179 2010-03-03 24 -364 val_364 2010-03-03 24 -87 val_87 2010-03-03 24 -490 val_490 2010-03-03 24 +172 val_172 2010-03-03 24 +42 val_42 2010-03-03 24 +129 val_129 2010-03-03 24 +158 val_158 2010-03-03 24 +119 val_119 2010-03-03 24 +496 val_496 2010-03-03 24 0 val_0 2010-03-03 24 -77 val_77 2010-03-03 24 -85 val_85 2010-03-03 24 -413 val_413 2010-03-03 24 -169 val_169 2010-03-03 24 -316 val_316 2010-03-03 24 -369 val_369 2010-03-03 24 -137 val_137 2010-03-03 24 -470 val_470 2010-03-03 24 -409 val_409 2010-03-03 24 -318 val_318 2010-03-03 24 -318 val_318 2010-03-03 24 -187 val_187 2010-03-03 24 +322 val_322 2010-03-03 24 197 val_197 2010-03-03 24 -282 val_282 2010-03-03 24 -98 val_98 2010-03-03 24 -457 val_457 2010-03-03 24 -481 val_481 2010-03-03 24 -288 val_288 2010-03-03 24 -196 val_196 2010-03-03 24 -95 val_95 2010-03-03 24 -308 val_308 2010-03-03 24 468 val_468 2010-03-03 24 -469 val_469 2010-03-03 24 -156 val_156 2010-03-03 24 -436 val_436 2010-03-03 24 -43 val_43 2010-03-03 24 -404 val_404 2010-03-03 24 -51 val_51 2010-03-03 24 -131 val_131 2010-03-03 24 -120 val_120 2010-03-03 24 -205 val_205 2010-03-03 24 -230 val_230 2010-03-03 24 -327 val_327 2010-03-03 24 -165 val_165 2010-03-03 24 -26 val_26 2010-03-03 24 -96 val_96 2010-03-03 24 -418 val_418 2010-03-03 24 -191 val_191 2010-03-03 24 -199 val_199 2010-03-03 24 -298 val_298 2010-03-03 24 -100 val_100 2010-03-03 24 -454 val_454 2010-03-03 24 393 val_393 2010-03-03 24 +454 val_454 2010-03-03 24 +100 val_100 2010-03-03 24 +298 val_298 2010-03-03 24 +199 val_199 2010-03-03 24 +191 val_191 2010-03-03 24 +418 val_418 2010-03-03 24 +96 val_96 2010-03-03 24 +26 val_26 2010-03-03 24 +165 val_165 2010-03-03 24 +327 val_327 2010-03-03 24 +230 val_230 2010-03-03 24 +205 val_205 2010-03-03 24 +120 val_120 2010-03-03 24 +131 val_131 2010-03-03 24 +51 val_51 2010-03-03 24 +404 val_404 2010-03-03 24 +43 val_43 2010-03-03 24 +436 val_436 2010-03-03 24 +156 val_156 2010-03-03 24 +469 val_469 2010-03-03 24 468 val_468 2010-03-03 24 +308 val_308 2010-03-03 24 +95 val_95 2010-03-03 24 +196 val_196 2010-03-03 24 +288 val_288 2010-03-03 24 +481 val_481 2010-03-03 24 +457 val_457 2010-03-03 24 +98 val_98 2010-03-03 24 +282 val_282 2010-03-03 24 197 val_197 2010-03-03 24 -322 val_322 2010-03-03 24 +187 val_187 2010-03-03 24 +318 val_318 2010-03-03 24 +318 val_318 2010-03-03 24 +409 val_409 2010-03-03 24 +470 val_470 2010-03-03 24 +137 val_137 2010-03-03 24 +369 val_369 2010-03-03 24 +316 val_316 2010-03-03 24 +169 val_169 2010-03-03 24 +413 val_413 2010-03-03 24 +85 val_85 2010-03-03 24 +77 val_77 2010-03-03 24 0 val_0 2010-03-03 24 -496 val_496 2010-03-03 24 -119 val_119 2010-03-03 24 -158 val_158 2010-03-03 24 -129 val_129 2010-03-03 24 -42 val_42 2010-03-03 24 -172 val_172 2010-03-03 24 +490 val_490 2010-03-03 24 +87 val_87 2010-03-03 24 +364 val_364 2010-03-03 24 179 val_179 2010-03-03 24 -431 val_431 2010-03-03 24 -463 val_463 2010-03-03 24 -24 val_24 2010-03-03 24 -348 val_348 2010-03-03 24 -208 val_208 2010-03-03 24 -230 val_230 2010-03-03 24 -411 val_411 2010-03-03 24 -8 val_8 2010-03-03 24 -58 val_58 2010-03-03 24 -466 val_466 2010-03-03 24 -348 val_348 2010-03-03 24 -84 val_84 2010-03-03 24 -217 val_217 2010-03-03 24 +118 val_118 2010-03-03 24 +134 val_134 2010-03-03 24 +395 val_395 2010-03-03 24 +282 val_282 2010-03-03 24 138 val_138 2010-03-03 24 +238 val_238 2010-03-03 24 +419 val_419 2010-03-03 24 +15 val_15 2010-03-03 24 +118 val_118 2010-03-03 24 +72 val_72 2010-03-03 24 +90 val_90 2010-03-03 24 +307 val_307 2010-03-03 24 +19 val_19 2010-03-03 24 +435 val_435 2010-03-03 24 +10 val_10 2010-03-03 24 +277 val_277 2010-03-03 24 +273 val_273 2010-03-03 24 +306 val_306 2010-03-03 24 +224 val_224 2010-03-03 24 +309 val_309 2010-03-03 24 +389 val_389 2010-03-03 24 +327 val_327 2010-03-03 24 +242 val_242 2010-03-03 24 +369 val_369 2010-03-03 24 +392 val_392 2010-03-03 24 272 val_272 2010-03-03 24 -373 val_373 2010-03-03 24 -353 val_353 2010-03-03 24 +331 val_331 2010-03-03 24 +401 val_401 2010-03-03 24 +242 val_242 2010-03-03 24 +452 val_452 2010-03-03 24 +177 val_177 2010-03-03 24 +226 val_226 2010-03-03 24 +5 val_5 2010-03-03 24 +497 val_497 2010-03-03 24 +402 val_402 2010-03-03 24 +396 val_396 2010-03-03 24 +317 val_317 2010-03-03 24 +395 val_395 2010-03-03 24 +58 val_58 2010-03-03 24 +35 val_35 2010-03-03 24 +336 val_336 2010-03-03 24 +95 val_95 2010-03-03 24 +11 val_11 2010-03-03 24 +168 val_168 2010-03-03 24 +34 val_34 2010-03-03 24 +229 val_229 2010-03-03 24 +233 val_233 2010-03-03 24 +143 val_143 2010-03-03 24 +472 val_472 2010-03-03 24 +322 val_322 2010-03-03 24 +498 val_498 2010-03-03 24 +160 val_160 2010-03-03 24 +195 val_195 2010-03-03 24 +42 val_42 2010-03-03 24 +321 val_321 2010-03-03 24 +430 val_430 2010-03-03 24 +119 val_119 2010-03-03 24 489 val_489 2010-03-03 24 -384 val_384 2010-03-03 24 -404 val_404 2010-03-03 24 -260 val_260 2010-03-03 24 -67 val_67 2010-03-03 24 -181 val_181 2010-03-03 24 +458 val_458 2010-03-03 24 +78 val_78 2010-03-03 24 +76 val_76 2010-03-03 24 +41 val_41 2010-03-03 24 +223 val_223 2010-03-03 24 +492 val_492 2010-03-03 24 +149 val_149 2010-03-03 24 +449 val_449 2010-03-03 24 +218 val_218 2010-03-03 24 +228 val_228 2010-03-03 24 +138 val_138 2010-03-03 24 +453 val_453 2010-03-03 24 +30 val_30 2010-03-03 24 +209 val_209 2010-03-03 24 +64 val_64 2010-03-03 24 +468 val_468 2010-03-03 24 +76 val_76 2010-03-03 24 +74 val_74 2010-03-03 24 +342 val_342 2010-03-03 24 +69 val_69 2010-03-03 24 230 val_230 2010-03-03 24 -12 val_12 2010-03-03 24 -284 val_284 2010-03-03 24 -180 val_180 2010-03-03 24 -333 val_333 2010-03-03 24 -83 val_83 2010-03-03 24 -241 val_241 2010-03-03 24 -137 val_137 2010-03-03 24 -275 val_275 2010-03-03 24 -311 val_311 2010-03-03 24 -332 val_332 2010-03-03 24 -318 val_318 2010-03-03 24 -65 val_65 2010-03-03 24 -221 val_221 2010-03-03 24 -289 val_289 2010-03-03 24 -176 val_176 2010-03-03 24 -278 val_278 2010-03-03 24 -430 val_430 2010-03-03 24 +33 val_33 2010-03-03 24 +368 val_368 2010-03-03 24 +103 val_103 2010-03-03 24 +296 val_296 2010-03-03 24 +113 val_113 2010-03-03 24 216 val_216 2010-03-03 24 -213 val_213 2010-03-03 24 +367 val_367 2010-03-03 24 +344 val_344 2010-03-03 24 +167 val_167 2010-03-03 24 +274 val_274 2010-03-03 24 +219 val_219 2010-03-03 24 239 val_239 2010-03-03 24 -103 val_103 2010-03-03 24 -138 val_138 2010-03-03 24 -51 val_51 2010-03-03 24 -459 val_459 2010-03-03 24 -54 val_54 2010-03-03 24 -176 val_176 2010-03-03 24 -187 val_187 2010-03-03 24 -286 val_286 2010-03-03 24 -192 val_192 2010-03-03 24 +485 val_485 2010-03-03 24 +116 val_116 2010-03-03 24 +223 val_223 2010-03-03 24 +256 val_256 2010-03-03 24 +263 val_263 2010-03-03 24 +70 val_70 2010-03-03 24 +487 val_487 2010-03-03 24 +480 val_480 2010-03-03 24 +401 val_401 2010-03-03 24 +288 val_288 2010-03-03 24 +191 val_191 2010-03-03 24 +5 val_5 2010-03-03 24 +244 val_244 2010-03-03 24 +438 val_438 2010-03-03 24 +128 val_128 2010-03-03 24 +467 val_467 2010-03-03 24 +432 val_432 2010-03-03 24 +202 val_202 2010-03-03 24 +316 val_316 2010-03-03 24 +229 val_229 2010-03-03 24 469 val_469 2010-03-03 24 -437 val_437 2010-03-03 24 -386 val_386 2010-03-03 24 +463 val_463 2010-03-03 24 +280 val_280 2010-03-03 24 +2 val_2 2010-03-03 24 +35 val_35 2010-03-03 24 +283 val_283 2010-03-03 24 +331 val_331 2010-03-03 24 +235 val_235 2010-03-03 24 +80 val_80 2010-03-03 24 +44 val_44 2010-03-03 24 +193 val_193 2010-03-03 24 +321 val_321 2010-03-03 24 +335 val_335 2010-03-03 24 +104 val_104 2010-03-03 24 +466 val_466 2010-03-03 24 +366 val_366 2010-03-03 24 +175 val_175 2010-03-03 24 +403 val_403 2010-03-03 24 +483 val_483 2010-03-03 24 +53 val_53 2010-03-03 24 +105 val_105 2010-03-03 24 +257 val_257 2010-03-03 24 +406 val_406 2010-03-03 24 +409 val_409 2010-03-03 24 +190 val_190 2010-03-03 24 +406 val_406 2010-03-03 24 +401 val_401 2010-03-03 24 +114 val_114 2010-03-03 24 +258 val_258 2010-03-03 24 +90 val_90 2010-03-03 24 +203 val_203 2010-03-03 24 +262 val_262 2010-03-03 24 +348 val_348 2010-03-03 24 +424 val_424 2010-03-03 24 +12 val_12 2010-03-03 24 +396 val_396 2010-03-03 24 +201 val_201 2010-03-03 24 +217 val_217 2010-03-03 24 +164 val_164 2010-03-03 24 +431 val_431 2010-03-03 24 +454 val_454 2010-03-03 24 +478 val_478 2010-03-03 24 +298 val_298 2010-03-03 24 125 val_125 2010-03-03 24 -498 val_498 2010-03-03 24 +431 val_431 2010-03-03 24 +164 val_164 2010-03-03 24 +424 val_424 2010-03-03 24 +187 val_187 2010-03-03 24 382 val_382 2010-03-03 24 -169 val_169 2010-03-03 24 -399 val_399 2010-03-03 24 -356 val_356 2010-03-03 24 -208 val_208 2010-03-03 24 -277 val_277 2010-03-03 24 -427 val_427 2010-03-03 24 -35 val_35 2010-03-03 24 -280 val_280 2010-03-03 24 -4 val_4 2010-03-03 24 -72 val_72 2010-03-03 24 -47 val_47 2010-03-03 24 -111 val_111 2010-03-03 24 -92 val_92 2010-03-03 24 -221 val_221 2010-03-03 24 -378 val_378 2010-03-03 24 -157 val_157 2010-03-03 24 -489 val_489 2010-03-03 24 -20 val_20 2010-03-03 24 -170 val_170 2010-03-03 24 -129 val_129 2010-03-03 24 -345 val_345 2010-03-03 24 +5 val_5 2010-03-03 24 +70 val_70 2010-03-03 24 +397 val_397 2010-03-03 24 +480 val_480 2010-03-03 24 +291 val_291 2010-03-03 24 +24 val_24 2010-03-03 24 +351 val_351 2010-03-03 24 +255 val_255 2010-03-03 24 +104 val_104 2010-03-03 24 +70 val_70 2010-03-03 24 +163 val_163 2010-03-03 24 438 val_438 2010-03-03 24 -149 val_149 2010-03-03 24 -205 val_205 2010-03-03 24 -302 val_302 2010-03-03 24 -57 val_57 2010-03-03 24 -316 val_316 2010-03-03 24 -311 val_311 2010-03-03 24 -128 val_128 2010-03-03 24 -455 val_455 2010-03-03 24 -0 val_0 2010-03-03 24 -339 val_339 2010-03-03 24 -203 val_203 2010-03-03 24 -155 val_155 2010-03-03 24 -113 val_113 2010-03-03 24 -17 val_17 2010-03-03 24 -475 val_475 2010-03-03 24 -195 val_195 2010-03-03 24 -167 val_167 2010-03-03 24 -325 val_325 2010-03-03 24 -367 val_367 2010-03-03 24 -342 val_342 2010-03-03 24 +119 val_119 2010-03-03 24 +414 val_414 2010-03-03 24 +200 val_200 2010-03-03 24 +491 val_491 2010-03-03 24 +237 val_237 2010-03-03 24 439 val_439 2010-03-03 24 -266 val_266 2010-03-03 24 -365 val_365 2010-03-03 24 -309 val_309 2010-03-03 24 -397 val_397 2010-03-03 24 -377 val_377 2010-03-03 24 -162 val_162 2010-03-03 24 -489 val_489 2010-03-03 24 +360 val_360 2010-03-03 24 +248 val_248 2010-03-03 24 +479 val_479 2010-03-03 24 +305 val_305 2010-03-03 24 417 val_417 2010-03-03 24 -247 val_247 2010-03-03 24 -396 val_396 2010-03-03 24 -399 val_399 2010-03-03 24 -174 val_174 2010-03-03 24 -208 val_208 2010-03-03 24 -466 val_466 2010-03-03 24 199 val_199 2010-03-03 24 -207 val_207 2010-03-03 24 -494 val_494 2010-03-03 24 -413 val_413 2010-03-03 24 -174 val_174 2010-03-03 24 -482 val_482 2010-03-03 24 -237 val_237 2010-03-03 24 -394 val_394 2010-03-03 24 -459 val_459 2010-03-03 24 -446 val_446 2010-03-03 24 -338 val_338 2010-03-03 24 -193 val_193 2010-03-03 24 -153 val_153 2010-03-03 24 -287 val_287 2010-03-03 24 -219 val_219 2010-03-03 24 -292 val_292 2010-03-03 24 -252 val_252 2010-03-03 24 -430 val_430 2010-03-03 24 -417 val_417 2010-03-03 24 -166 val_166 2010-03-03 24 -403 val_403 2010-03-03 24 -82 val_82 2010-03-03 24 -15 val_15 2010-03-03 24 -209 val_209 2010-03-03 24 -277 val_277 2010-03-03 24 -281 val_281 2010-03-03 24 -327 val_327 2010-03-03 24 -37 val_37 2010-03-03 24 -495 val_495 2010-03-03 24 -145 val_145 2010-03-03 24 -469 val_469 2010-03-03 24 -152 val_152 2010-03-03 24 -374 val_374 2010-03-03 24 +444 val_444 2010-03-03 24 +120 val_120 2010-03-03 24 429 val_429 2010-03-03 24 +169 val_169 2010-03-03 24 +443 val_443 2010-03-03 24 +323 val_323 2010-03-03 24 +325 val_325 2010-03-03 24 +277 val_277 2010-03-03 24 +230 val_230 2010-03-03 24 +478 val_478 2010-03-03 24 +178 val_178 2010-03-03 24 +468 val_468 2010-03-03 24 +310 val_310 2010-03-03 24 +317 val_317 2010-03-03 24 +333 val_333 2010-03-03 24 +493 val_493 2010-03-03 24 +460 val_460 2010-03-03 24 +207 val_207 2010-03-03 24 +249 val_249 2010-03-03 24 +265 val_265 2010-03-03 24 +480 val_480 2010-03-03 24 +83 val_83 2010-03-03 24 +136 val_136 2010-03-03 24 +353 val_353 2010-03-03 24 +172 val_172 2010-03-03 24 +214 val_214 2010-03-03 24 +462 val_462 2010-03-03 24 +233 val_233 2010-03-03 24 406 val_406 2010-03-03 24 +133 val_133 2010-03-03 24 +175 val_175 2010-03-03 24 +189 val_189 2010-03-03 24 +454 val_454 2010-03-03 24 +375 val_375 2010-03-03 24 +401 val_401 2010-03-03 24 +421 val_421 2010-03-03 24 +407 val_407 2010-03-03 24 +384 val_384 2010-03-03 24 +256 val_256 2010-03-03 24 +26 val_26 2010-03-03 24 +134 val_134 2010-03-03 24 +67 val_67 2010-03-03 24 +384 val_384 2010-03-03 24 +379 val_379 2010-03-03 24 +18 val_18 2010-03-03 24 +462 val_462 2010-03-03 24 +492 val_492 2010-03-03 24 +100 val_100 2010-03-03 24 +298 val_298 2010-03-03 24 +9 val_9 2010-03-03 24 +341 val_341 2010-03-03 24 +498 val_498 2010-03-03 24 146 val_146 2010-03-03 24 -213 val_213 2010-03-03 24 -128 val_128 2010-03-03 24 -66 val_66 2010-03-03 24 -369 val_369 2010-03-03 24 -224 val_224 2010-03-03 24 +458 val_458 2010-03-03 24 +362 val_362 2010-03-03 24 +186 val_186 2010-03-03 24 +285 val_285 2010-03-03 24 +348 val_348 2010-03-03 24 +167 val_167 2010-03-03 24 +18 val_18 2010-03-03 24 273 val_273 2010-03-03 24 -150 val_150 2010-03-03 24 -401 val_401 2010-03-03 24 -193 val_193 2010-03-03 24 -265 val_265 2010-03-03 24 -484 val_484 2010-03-03 24 -98 val_98 2010-03-03 24 -278 val_278 2010-03-03 24 -255 val_255 2010-03-03 24 -409 val_409 2010-03-03 24 -165 val_165 2010-03-03 24 -27 val_27 2010-03-03 24 -311 val_311 2010-03-03 24 -86 val_86 2010-03-03 24 -238 val_238 2010-03-03 24 +183 val_183 2010-03-03 24 +281 val_281 2010-03-03 24 +344 val_344 2010-03-03 24 +97 val_97 2010-03-03 24 +469 val_469 2010-03-03 24 +315 val_315 2010-03-03 24 +84 val_84 2010-03-03 24 +28 val_28 2010-03-03 24 +37 val_37 2010-03-03 24 +448 val_448 2010-03-03 24 +152 val_152 2010-03-03 24 +348 val_348 2010-03-03 24 +307 val_307 2010-03-03 24 +194 val_194 2010-03-03 24 +414 val_414 2010-03-03 24 +477 val_477 2010-03-03 24 +222 val_222 2010-03-03 24 +126 val_126 2010-03-03 24 +90 val_90 2010-03-03 24 +169 val_169 2010-03-03 24 +403 val_403 2010-03-03 24 +400 val_400 2010-03-03 24 +200 val_200 2010-03-03 24 97 val_97 2010-03-03 24 diff --git ql/src/test/results/clientpositive/llap/load_dyn_part13.q.out ql/src/test/results/clientpositive/llap/load_dyn_part13.q.out new file mode 100644 index 0000000000..708720eb50 --- /dev/null +++ ql/src/test/results/clientpositive/llap/load_dyn_part13.q.out @@ -0,0 +1,288 @@ +PREHOOK: query: show partitions srcpart +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@srcpart +POSTHOOK: query: show partitions srcpart +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@srcpart +ds=2008-04-08/hr=11 +ds=2008-04-08/hr=12 +ds=2008-04-09/hr=11 +ds=2008-04-09/hr=12 +PREHOOK: query: create table if not exists nzhang_part13 like srcpart +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_part13 +POSTHOOK: query: create table if not exists nzhang_part13 like srcpart +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@nzhang_part13 +PREHOOK: query: describe extended nzhang_part13 +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@nzhang_part13 +POSTHOOK: query: describe extended nzhang_part13 +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@nzhang_part13 +key string default +value string default +ds string +hr string + +# Partition Information +# col_name data_type comment +ds string +hr string + +#### A masked pattern was here #### +PREHOOK: query: explain +insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) +select * from ( + select key, value, '22' + from src + where key < 20 + union all + select key, value, '33' + from src + where key > 20 and key < 40) s +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@nzhang_part13@ds=2010-03-03 +POSTHOOK: query: explain +insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) +select * from ( + select key, value, '22' + from src + where key < 20 + union all + select key, value, '33' + from src + where key > 20 and key < 40) s +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (UDFToDouble(key) < 20.0D) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) < 20.0D) (type: boolean) + Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), '22' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 166 Data size: 43824 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part13 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2010-03-03' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 221 Data size: 79118 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 40.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 40.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), '33' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 55 Data size: 14520 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part13 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2010-03-03' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 221 Data size: 79118 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 2 + Vertex: Union 2 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds 2010-03-03 + hr + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part13 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.nzhang_part13 + +PREHOOK: query: insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) +select * from ( + select key, value, '22' + from src + where key < 20 + union all + select key, value, '33' + from src + where key > 20 and key < 40) s +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@nzhang_part13@ds=2010-03-03 +POSTHOOK: query: insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) +select * from ( + select key, value, '22' + from src + where key < 20 + union all + select key, value, '33' + from src + where key > 20 and key < 40) s +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@nzhang_part13@ds=2010-03-03/hr=22 +POSTHOOK: Output: default@nzhang_part13@ds=2010-03-03/hr=33 +POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: show partitions nzhang_part13 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@nzhang_part13 +POSTHOOK: query: show partitions nzhang_part13 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@nzhang_part13 +ds=2010-03-03/hr=22 +ds=2010-03-03/hr=33 +PREHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null +PREHOOK: type: QUERY +PREHOOK: Input: default@nzhang_part13 +PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22 +PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33 +#### A masked pattern was here #### +POSTHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null +POSTHOOK: type: QUERY +POSTHOOK: Input: default@nzhang_part13 +POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22 +POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33 +#### A masked pattern was here #### +15 val_15 2010-03-03 22 +17 val_17 2010-03-03 22 +0 val_0 2010-03-03 22 +4 val_4 2010-03-03 22 +12 val_12 2010-03-03 22 +8 val_8 2010-03-03 22 +0 val_0 2010-03-03 22 +0 val_0 2010-03-03 22 +15 val_15 2010-03-03 22 +19 val_19 2010-03-03 22 +10 val_10 2010-03-03 22 +5 val_5 2010-03-03 22 +11 val_11 2010-03-03 22 +5 val_5 2010-03-03 22 +2 val_2 2010-03-03 22 +12 val_12 2010-03-03 22 +5 val_5 2010-03-03 22 +18 val_18 2010-03-03 22 +9 val_9 2010-03-03 22 +18 val_18 2010-03-03 22 +27 val_27 2010-03-03 33 +37 val_37 2010-03-03 33 +35 val_35 2010-03-03 33 +24 val_24 2010-03-03 33 +26 val_26 2010-03-03 33 +35 val_35 2010-03-03 33 +34 val_34 2010-03-03 33 +30 val_30 2010-03-03 33 +33 val_33 2010-03-03 33 +35 val_35 2010-03-03 33 +24 val_24 2010-03-03 33 +26 val_26 2010-03-03 33 +28 val_28 2010-03-03 33 +37 val_37 2010-03-03 33 diff --git ql/src/test/results/clientpositive/llap/load_dyn_part14.q.out ql/src/test/results/clientpositive/llap/load_dyn_part14.q.out new file mode 100644 index 0000000000..cf08318e71 --- /dev/null +++ ql/src/test/results/clientpositive/llap/load_dyn_part14.q.out @@ -0,0 +1,320 @@ +PREHOOK: query: create table if not exists nzhang_part14_n0 (key string) + partitioned by (value string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@nzhang_part14_n0 +POSTHOOK: query: create table if not exists nzhang_part14_n0 (key string) + partitioned by (value string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@nzhang_part14_n0 +PREHOOK: query: describe extended nzhang_part14_n0 +PREHOOK: type: DESCTABLE +PREHOOK: Input: default@nzhang_part14_n0 +POSTHOOK: query: describe extended nzhang_part14_n0 +POSTHOOK: type: DESCTABLE +POSTHOOK: Input: default@nzhang_part14_n0 +key string +value string + +# Partition Information +# col_name data_type comment +value string + +#### A masked pattern was here #### +PREHOOK: query: explain +insert overwrite table nzhang_part14_n0 partition(value) +select key, value from ( + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a + union all + select * from (select 'k2' as key, '' as value from src limit 2)b + union all + select * from (select 'k3' as key, ' ' as value from src limit 2)c +) T +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@nzhang_part14_n0 +POSTHOOK: query: explain +insert overwrite table nzhang_part14_n0 partition(value) +select key, value from ( + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a + union all + select * from (select 'k2' as key, '' as value from src limit 2)b + union all + select * from (select 'k3' as key, ' ' as value from src limit 2)c +) T +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) + Reducer 4 <- Union 3 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) + Reducer 8 <- Map 7 (CUSTOM_SIMPLE_EDGE), Union 3 (CONTAINS) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'k1' (type: string), null (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 256 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + keys: value (type: string) + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'k2' (type: string), '' (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + keys: value (type: string) + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct) + Reducer 8 + Execution mode: llap + Reduce Operator Tree: + Limit + Number of rows: 2 + Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: 'k3' (type: string), ' ' (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14_n0 + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: key, value + Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll') + keys: value (type: string) + minReductionHashAggr: 0.6666666 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: struct) + Union 3 + Vertex: Union 3 + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + value + replace: true + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14_n0 + + Stage: Stage-3 + Stats Work + Basic Stats Work: + Column Stats Desc: + Columns: key + Column Types: string + Table: default.nzhang_part14_n0 + +PREHOOK: query: insert overwrite table nzhang_part14_n0 partition(value) +select key, value from ( + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a + union all + select * from (select 'k2' as key, '' as value from src limit 2)b + union all + select * from (select 'k3' as key, ' ' as value from src limit 2)c +) T +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Output: default@nzhang_part14_n0 +POSTHOOK: query: insert overwrite table nzhang_part14_n0 partition(value) +select key, value from ( + select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a + union all + select * from (select 'k2' as key, '' as value from src limit 2)b + union all + select * from (select 'k3' as key, ' ' as value from src limit 2)c +) T +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Output: default@nzhang_part14_n0@value= +POSTHOOK: Output: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ +POSTHOOK: Lineage: nzhang_part14_n0 PARTITION(value= ).key EXPRESSION [] +POSTHOOK: Lineage: nzhang_part14_n0 PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION [] +PREHOOK: query: show partitions nzhang_part14_n0 +PREHOOK: type: SHOWPARTITIONS +PREHOOK: Input: default@nzhang_part14_n0 +POSTHOOK: query: show partitions nzhang_part14_n0 +POSTHOOK: type: SHOWPARTITIONS +POSTHOOK: Input: default@nzhang_part14_n0 +value= +value=__HIVE_DEFAULT_PARTITION__ +PREHOOK: query: select * from nzhang_part14_n0 where value <> 'a' +PREHOOK: type: QUERY +PREHOOK: Input: default@nzhang_part14_n0 +PREHOOK: Input: default@nzhang_part14_n0@value= +PREHOOK: Input: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ +#### A masked pattern was here #### +POSTHOOK: query: select * from nzhang_part14_n0 where value <> 'a' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@nzhang_part14_n0 +POSTHOOK: Input: default@nzhang_part14_n0@value= +POSTHOOK: Input: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ +#### A masked pattern was here #### +k1 __HIVE_DEFAULT_PARTITION__ +k1 __HIVE_DEFAULT_PARTITION__ +k2 __HIVE_DEFAULT_PARTITION__ +k2 __HIVE_DEFAULT_PARTITION__ +k3 +k3 diff --git ql/src/test/results/clientpositive/load_dyn_part4.q.out ql/src/test/results/clientpositive/llap/load_dyn_part4.q.out similarity index 96% rename from ql/src/test/results/clientpositive/load_dyn_part4.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part4.q.out index 598cab96ec..48ca6a59f3 100644 --- ql/src/test/results/clientpositive/load_dyn_part4.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part4.q.out @@ -62,65 +62,79 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds is not null and hr is not null) (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part4 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds is not null and hr is not null) (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part4 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 4 Data size: 4992 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -135,7 +149,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part4 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: diff --git ql/src/test/results/clientpositive/load_dyn_part6.q.out ql/src/test/results/clientpositive/llap/load_dyn_part6.q.out similarity index 100% rename from ql/src/test/results/clientpositive/load_dyn_part6.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part6.q.out index a0f574d5de..993f42f2f1 100644 --- ql/src/test/results/clientpositive/load_dyn_part6.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part6.q.out @@ -65,1005 +65,1005 @@ POSTHOOK: Input: default@nzhang_part6 POSTHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=11 #### A masked pattern was here #### 238 val_238 2010-03-03 11 -97 val_97 2010-03-03 11 -200 val_200 2010-03-03 11 -400 val_400 2010-03-03 11 -403 val_403 2010-03-03 11 -169 val_169 2010-03-03 11 -90 val_90 2010-03-03 11 -126 val_126 2010-03-03 11 -222 val_222 2010-03-03 11 -477 val_477 2010-03-03 11 -414 val_414 2010-03-03 11 -194 val_194 2010-03-03 11 -307 val_307 2010-03-03 11 -348 val_348 2010-03-03 11 -152 val_152 2010-03-03 11 -448 val_448 2010-03-03 11 -37 val_37 2010-03-03 11 -28 val_28 2010-03-03 11 -84 val_84 2010-03-03 11 -315 val_315 2010-03-03 11 -469 val_469 2010-03-03 11 -97 val_97 2010-03-03 11 -344 val_344 2010-03-03 11 -281 val_281 2010-03-03 11 -183 val_183 2010-03-03 11 +86 val_86 2010-03-03 11 +311 val_311 2010-03-03 11 +27 val_27 2010-03-03 11 +165 val_165 2010-03-03 11 +409 val_409 2010-03-03 11 +255 val_255 2010-03-03 11 +278 val_278 2010-03-03 11 +98 val_98 2010-03-03 11 +484 val_484 2010-03-03 11 +265 val_265 2010-03-03 11 +193 val_193 2010-03-03 11 +401 val_401 2010-03-03 11 +150 val_150 2010-03-03 11 273 val_273 2010-03-03 11 -18 val_18 2010-03-03 11 -167 val_167 2010-03-03 11 -348 val_348 2010-03-03 11 -285 val_285 2010-03-03 11 -186 val_186 2010-03-03 11 -362 val_362 2010-03-03 11 -458 val_458 2010-03-03 11 +224 val_224 2010-03-03 11 +369 val_369 2010-03-03 11 +66 val_66 2010-03-03 11 +128 val_128 2010-03-03 11 +213 val_213 2010-03-03 11 146 val_146 2010-03-03 11 -498 val_498 2010-03-03 11 -341 val_341 2010-03-03 11 -9 val_9 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -492 val_492 2010-03-03 11 -462 val_462 2010-03-03 11 -18 val_18 2010-03-03 11 -379 val_379 2010-03-03 11 -384 val_384 2010-03-03 11 -67 val_67 2010-03-03 11 -134 val_134 2010-03-03 11 -26 val_26 2010-03-03 11 -256 val_256 2010-03-03 11 -384 val_384 2010-03-03 11 -407 val_407 2010-03-03 11 -421 val_421 2010-03-03 11 -401 val_401 2010-03-03 11 -375 val_375 2010-03-03 11 -454 val_454 2010-03-03 11 -189 val_189 2010-03-03 11 -175 val_175 2010-03-03 11 -133 val_133 2010-03-03 11 406 val_406 2010-03-03 11 -233 val_233 2010-03-03 11 -462 val_462 2010-03-03 11 -214 val_214 2010-03-03 11 -172 val_172 2010-03-03 11 -353 val_353 2010-03-03 11 -136 val_136 2010-03-03 11 -83 val_83 2010-03-03 11 -480 val_480 2010-03-03 11 -265 val_265 2010-03-03 11 -249 val_249 2010-03-03 11 -207 val_207 2010-03-03 11 -460 val_460 2010-03-03 11 -493 val_493 2010-03-03 11 -333 val_333 2010-03-03 11 -317 val_317 2010-03-03 11 -310 val_310 2010-03-03 11 -468 val_468 2010-03-03 11 -178 val_178 2010-03-03 11 -478 val_478 2010-03-03 11 -230 val_230 2010-03-03 11 -277 val_277 2010-03-03 11 -325 val_325 2010-03-03 11 -323 val_323 2010-03-03 11 -443 val_443 2010-03-03 11 -169 val_169 2010-03-03 11 429 val_429 2010-03-03 11 -120 val_120 2010-03-03 11 -444 val_444 2010-03-03 11 +374 val_374 2010-03-03 11 +152 val_152 2010-03-03 11 +469 val_469 2010-03-03 11 +145 val_145 2010-03-03 11 +495 val_495 2010-03-03 11 +37 val_37 2010-03-03 11 +327 val_327 2010-03-03 11 +281 val_281 2010-03-03 11 +277 val_277 2010-03-03 11 +209 val_209 2010-03-03 11 +15 val_15 2010-03-03 11 +82 val_82 2010-03-03 11 +403 val_403 2010-03-03 11 +166 val_166 2010-03-03 11 +417 val_417 2010-03-03 11 +430 val_430 2010-03-03 11 +252 val_252 2010-03-03 11 +292 val_292 2010-03-03 11 +219 val_219 2010-03-03 11 +287 val_287 2010-03-03 11 +153 val_153 2010-03-03 11 +193 val_193 2010-03-03 11 +338 val_338 2010-03-03 11 +446 val_446 2010-03-03 11 +459 val_459 2010-03-03 11 +394 val_394 2010-03-03 11 +237 val_237 2010-03-03 11 +482 val_482 2010-03-03 11 +174 val_174 2010-03-03 11 +413 val_413 2010-03-03 11 +494 val_494 2010-03-03 11 +207 val_207 2010-03-03 11 199 val_199 2010-03-03 11 +466 val_466 2010-03-03 11 +208 val_208 2010-03-03 11 +174 val_174 2010-03-03 11 +399 val_399 2010-03-03 11 +396 val_396 2010-03-03 11 +247 val_247 2010-03-03 11 417 val_417 2010-03-03 11 -305 val_305 2010-03-03 11 -479 val_479 2010-03-03 11 -248 val_248 2010-03-03 11 -360 val_360 2010-03-03 11 +489 val_489 2010-03-03 11 +162 val_162 2010-03-03 11 +377 val_377 2010-03-03 11 +397 val_397 2010-03-03 11 +309 val_309 2010-03-03 11 +365 val_365 2010-03-03 11 +266 val_266 2010-03-03 11 439 val_439 2010-03-03 11 -237 val_237 2010-03-03 11 -491 val_491 2010-03-03 11 -200 val_200 2010-03-03 11 -414 val_414 2010-03-03 11 -119 val_119 2010-03-03 11 +342 val_342 2010-03-03 11 +367 val_367 2010-03-03 11 +325 val_325 2010-03-03 11 +167 val_167 2010-03-03 11 +195 val_195 2010-03-03 11 +475 val_475 2010-03-03 11 +17 val_17 2010-03-03 11 +113 val_113 2010-03-03 11 +155 val_155 2010-03-03 11 +203 val_203 2010-03-03 11 +339 val_339 2010-03-03 11 +0 val_0 2010-03-03 11 +455 val_455 2010-03-03 11 +128 val_128 2010-03-03 11 +311 val_311 2010-03-03 11 +316 val_316 2010-03-03 11 +57 val_57 2010-03-03 11 +302 val_302 2010-03-03 11 +205 val_205 2010-03-03 11 +149 val_149 2010-03-03 11 438 val_438 2010-03-03 11 -163 val_163 2010-03-03 11 -70 val_70 2010-03-03 11 -104 val_104 2010-03-03 11 -255 val_255 2010-03-03 11 -351 val_351 2010-03-03 11 -24 val_24 2010-03-03 11 -291 val_291 2010-03-03 11 -480 val_480 2010-03-03 11 -397 val_397 2010-03-03 11 -70 val_70 2010-03-03 11 -5 val_5 2010-03-03 11 +345 val_345 2010-03-03 11 +129 val_129 2010-03-03 11 +170 val_170 2010-03-03 11 +20 val_20 2010-03-03 11 +489 val_489 2010-03-03 11 +157 val_157 2010-03-03 11 +378 val_378 2010-03-03 11 +221 val_221 2010-03-03 11 +92 val_92 2010-03-03 11 +111 val_111 2010-03-03 11 +47 val_47 2010-03-03 11 +72 val_72 2010-03-03 11 +4 val_4 2010-03-03 11 +280 val_280 2010-03-03 11 +35 val_35 2010-03-03 11 +427 val_427 2010-03-03 11 +277 val_277 2010-03-03 11 +208 val_208 2010-03-03 11 +356 val_356 2010-03-03 11 +399 val_399 2010-03-03 11 +169 val_169 2010-03-03 11 382 val_382 2010-03-03 11 -187 val_187 2010-03-03 11 -424 val_424 2010-03-03 11 -164 val_164 2010-03-03 11 -431 val_431 2010-03-03 11 +498 val_498 2010-03-03 11 125 val_125 2010-03-03 11 -298 val_298 2010-03-03 11 -478 val_478 2010-03-03 11 -454 val_454 2010-03-03 11 -431 val_431 2010-03-03 11 -164 val_164 2010-03-03 11 -217 val_217 2010-03-03 11 -201 val_201 2010-03-03 11 -396 val_396 2010-03-03 11 -12 val_12 2010-03-03 11 -424 val_424 2010-03-03 11 -348 val_348 2010-03-03 11 -262 val_262 2010-03-03 11 -203 val_203 2010-03-03 11 -90 val_90 2010-03-03 11 -258 val_258 2010-03-03 11 -114 val_114 2010-03-03 11 -401 val_401 2010-03-03 11 -406 val_406 2010-03-03 11 -190 val_190 2010-03-03 11 -409 val_409 2010-03-03 11 -406 val_406 2010-03-03 11 -257 val_257 2010-03-03 11 -105 val_105 2010-03-03 11 -53 val_53 2010-03-03 11 -483 val_483 2010-03-03 11 -403 val_403 2010-03-03 11 -175 val_175 2010-03-03 11 -366 val_366 2010-03-03 11 -466 val_466 2010-03-03 11 -104 val_104 2010-03-03 11 -335 val_335 2010-03-03 11 -321 val_321 2010-03-03 11 -193 val_193 2010-03-03 11 -44 val_44 2010-03-03 11 -80 val_80 2010-03-03 11 -235 val_235 2010-03-03 11 -331 val_331 2010-03-03 11 -283 val_283 2010-03-03 11 -35 val_35 2010-03-03 11 -2 val_2 2010-03-03 11 -280 val_280 2010-03-03 11 -463 val_463 2010-03-03 11 +386 val_386 2010-03-03 11 +437 val_437 2010-03-03 11 469 val_469 2010-03-03 11 -229 val_229 2010-03-03 11 -316 val_316 2010-03-03 11 -202 val_202 2010-03-03 11 -432 val_432 2010-03-03 11 -467 val_467 2010-03-03 11 -128 val_128 2010-03-03 11 -438 val_438 2010-03-03 11 -244 val_244 2010-03-03 11 -5 val_5 2010-03-03 11 -191 val_191 2010-03-03 11 -288 val_288 2010-03-03 11 -401 val_401 2010-03-03 11 -480 val_480 2010-03-03 11 -487 val_487 2010-03-03 11 -70 val_70 2010-03-03 11 -263 val_263 2010-03-03 11 -256 val_256 2010-03-03 11 -223 val_223 2010-03-03 11 -116 val_116 2010-03-03 11 -485 val_485 2010-03-03 11 +192 val_192 2010-03-03 11 +286 val_286 2010-03-03 11 +187 val_187 2010-03-03 11 +176 val_176 2010-03-03 11 +54 val_54 2010-03-03 11 +459 val_459 2010-03-03 11 +51 val_51 2010-03-03 11 +138 val_138 2010-03-03 11 +103 val_103 2010-03-03 11 239 val_239 2010-03-03 11 -219 val_219 2010-03-03 11 -274 val_274 2010-03-03 11 -167 val_167 2010-03-03 11 -344 val_344 2010-03-03 11 -367 val_367 2010-03-03 11 +213 val_213 2010-03-03 11 216 val_216 2010-03-03 11 -113 val_113 2010-03-03 11 -296 val_296 2010-03-03 11 -103 val_103 2010-03-03 11 -368 val_368 2010-03-03 11 -33 val_33 2010-03-03 11 +430 val_430 2010-03-03 11 +278 val_278 2010-03-03 11 +176 val_176 2010-03-03 11 +289 val_289 2010-03-03 11 +221 val_221 2010-03-03 11 +65 val_65 2010-03-03 11 +318 val_318 2010-03-03 11 +332 val_332 2010-03-03 11 +311 val_311 2010-03-03 11 +275 val_275 2010-03-03 11 +137 val_137 2010-03-03 11 +241 val_241 2010-03-03 11 +83 val_83 2010-03-03 11 +333 val_333 2010-03-03 11 +180 val_180 2010-03-03 11 +284 val_284 2010-03-03 11 +12 val_12 2010-03-03 11 230 val_230 2010-03-03 11 -69 val_69 2010-03-03 11 -342 val_342 2010-03-03 11 -74 val_74 2010-03-03 11 -76 val_76 2010-03-03 11 -468 val_468 2010-03-03 11 -64 val_64 2010-03-03 11 -209 val_209 2010-03-03 11 -30 val_30 2010-03-03 11 -453 val_453 2010-03-03 11 -138 val_138 2010-03-03 11 -228 val_228 2010-03-03 11 -218 val_218 2010-03-03 11 -449 val_449 2010-03-03 11 -149 val_149 2010-03-03 11 -492 val_492 2010-03-03 11 -223 val_223 2010-03-03 11 -41 val_41 2010-03-03 11 -76 val_76 2010-03-03 11 -78 val_78 2010-03-03 11 -458 val_458 2010-03-03 11 +181 val_181 2010-03-03 11 +67 val_67 2010-03-03 11 +260 val_260 2010-03-03 11 +404 val_404 2010-03-03 11 +384 val_384 2010-03-03 11 489 val_489 2010-03-03 11 -119 val_119 2010-03-03 11 -430 val_430 2010-03-03 11 -321 val_321 2010-03-03 11 -42 val_42 2010-03-03 11 -195 val_195 2010-03-03 11 -160 val_160 2010-03-03 11 -498 val_498 2010-03-03 11 -322 val_322 2010-03-03 11 -472 val_472 2010-03-03 11 -143 val_143 2010-03-03 11 -233 val_233 2010-03-03 11 -229 val_229 2010-03-03 11 -34 val_34 2010-03-03 11 -168 val_168 2010-03-03 11 -11 val_11 2010-03-03 11 -95 val_95 2010-03-03 11 -336 val_336 2010-03-03 11 -35 val_35 2010-03-03 11 -58 val_58 2010-03-03 11 -395 val_395 2010-03-03 11 -317 val_317 2010-03-03 11 -396 val_396 2010-03-03 11 -402 val_402 2010-03-03 11 -497 val_497 2010-03-03 11 -5 val_5 2010-03-03 11 -226 val_226 2010-03-03 11 -177 val_177 2010-03-03 11 -452 val_452 2010-03-03 11 -242 val_242 2010-03-03 11 -401 val_401 2010-03-03 11 -331 val_331 2010-03-03 11 +353 val_353 2010-03-03 11 +373 val_373 2010-03-03 11 272 val_272 2010-03-03 11 -392 val_392 2010-03-03 11 -369 val_369 2010-03-03 11 -242 val_242 2010-03-03 11 -327 val_327 2010-03-03 11 -389 val_389 2010-03-03 11 -309 val_309 2010-03-03 11 -224 val_224 2010-03-03 11 -306 val_306 2010-03-03 11 -273 val_273 2010-03-03 11 -277 val_277 2010-03-03 11 -10 val_10 2010-03-03 11 -435 val_435 2010-03-03 11 -19 val_19 2010-03-03 11 -307 val_307 2010-03-03 11 -90 val_90 2010-03-03 11 -72 val_72 2010-03-03 11 -118 val_118 2010-03-03 11 -15 val_15 2010-03-03 11 -419 val_419 2010-03-03 11 -238 val_238 2010-03-03 11 138 val_138 2010-03-03 11 -282 val_282 2010-03-03 11 -395 val_395 2010-03-03 11 -134 val_134 2010-03-03 11 -118 val_118 2010-03-03 11 +217 val_217 2010-03-03 11 +84 val_84 2010-03-03 11 +348 val_348 2010-03-03 11 +466 val_466 2010-03-03 11 +58 val_58 2010-03-03 11 +8 val_8 2010-03-03 11 +411 val_411 2010-03-03 11 +230 val_230 2010-03-03 11 +208 val_208 2010-03-03 11 +348 val_348 2010-03-03 11 +24 val_24 2010-03-03 11 +463 val_463 2010-03-03 11 +431 val_431 2010-03-03 11 179 val_179 2010-03-03 11 -364 val_364 2010-03-03 11 -87 val_87 2010-03-03 11 -490 val_490 2010-03-03 11 +172 val_172 2010-03-03 11 +42 val_42 2010-03-03 11 +129 val_129 2010-03-03 11 +158 val_158 2010-03-03 11 +119 val_119 2010-03-03 11 +496 val_496 2010-03-03 11 0 val_0 2010-03-03 11 -77 val_77 2010-03-03 11 -85 val_85 2010-03-03 11 -413 val_413 2010-03-03 11 -169 val_169 2010-03-03 11 -316 val_316 2010-03-03 11 -369 val_369 2010-03-03 11 -137 val_137 2010-03-03 11 -470 val_470 2010-03-03 11 -409 val_409 2010-03-03 11 -318 val_318 2010-03-03 11 -318 val_318 2010-03-03 11 -187 val_187 2010-03-03 11 +322 val_322 2010-03-03 11 197 val_197 2010-03-03 11 -282 val_282 2010-03-03 11 -98 val_98 2010-03-03 11 -457 val_457 2010-03-03 11 -481 val_481 2010-03-03 11 -288 val_288 2010-03-03 11 -196 val_196 2010-03-03 11 -95 val_95 2010-03-03 11 -308 val_308 2010-03-03 11 468 val_468 2010-03-03 11 -469 val_469 2010-03-03 11 -156 val_156 2010-03-03 11 -436 val_436 2010-03-03 11 -43 val_43 2010-03-03 11 -404 val_404 2010-03-03 11 -51 val_51 2010-03-03 11 -131 val_131 2010-03-03 11 -120 val_120 2010-03-03 11 -205 val_205 2010-03-03 11 -230 val_230 2010-03-03 11 -327 val_327 2010-03-03 11 -165 val_165 2010-03-03 11 -26 val_26 2010-03-03 11 -96 val_96 2010-03-03 11 -418 val_418 2010-03-03 11 -191 val_191 2010-03-03 11 -199 val_199 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -454 val_454 2010-03-03 11 393 val_393 2010-03-03 11 +454 val_454 2010-03-03 11 +100 val_100 2010-03-03 11 +298 val_298 2010-03-03 11 +199 val_199 2010-03-03 11 +191 val_191 2010-03-03 11 +418 val_418 2010-03-03 11 +96 val_96 2010-03-03 11 +26 val_26 2010-03-03 11 +165 val_165 2010-03-03 11 +327 val_327 2010-03-03 11 +230 val_230 2010-03-03 11 +205 val_205 2010-03-03 11 +120 val_120 2010-03-03 11 +131 val_131 2010-03-03 11 +51 val_51 2010-03-03 11 +404 val_404 2010-03-03 11 +43 val_43 2010-03-03 11 +436 val_436 2010-03-03 11 +156 val_156 2010-03-03 11 +469 val_469 2010-03-03 11 468 val_468 2010-03-03 11 +308 val_308 2010-03-03 11 +95 val_95 2010-03-03 11 +196 val_196 2010-03-03 11 +288 val_288 2010-03-03 11 +481 val_481 2010-03-03 11 +457 val_457 2010-03-03 11 +98 val_98 2010-03-03 11 +282 val_282 2010-03-03 11 197 val_197 2010-03-03 11 -322 val_322 2010-03-03 11 +187 val_187 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +409 val_409 2010-03-03 11 +470 val_470 2010-03-03 11 +137 val_137 2010-03-03 11 +369 val_369 2010-03-03 11 +316 val_316 2010-03-03 11 +169 val_169 2010-03-03 11 +413 val_413 2010-03-03 11 +85 val_85 2010-03-03 11 +77 val_77 2010-03-03 11 0 val_0 2010-03-03 11 -496 val_496 2010-03-03 11 -119 val_119 2010-03-03 11 -158 val_158 2010-03-03 11 -129 val_129 2010-03-03 11 -42 val_42 2010-03-03 11 -172 val_172 2010-03-03 11 +490 val_490 2010-03-03 11 +87 val_87 2010-03-03 11 +364 val_364 2010-03-03 11 179 val_179 2010-03-03 11 -431 val_431 2010-03-03 11 -463 val_463 2010-03-03 11 -24 val_24 2010-03-03 11 -348 val_348 2010-03-03 11 -208 val_208 2010-03-03 11 -230 val_230 2010-03-03 11 -411 val_411 2010-03-03 11 -8 val_8 2010-03-03 11 -58 val_58 2010-03-03 11 -466 val_466 2010-03-03 11 -348 val_348 2010-03-03 11 -84 val_84 2010-03-03 11 -217 val_217 2010-03-03 11 +118 val_118 2010-03-03 11 +134 val_134 2010-03-03 11 +395 val_395 2010-03-03 11 +282 val_282 2010-03-03 11 138 val_138 2010-03-03 11 +238 val_238 2010-03-03 11 +419 val_419 2010-03-03 11 +15 val_15 2010-03-03 11 +118 val_118 2010-03-03 11 +72 val_72 2010-03-03 11 +90 val_90 2010-03-03 11 +307 val_307 2010-03-03 11 +19 val_19 2010-03-03 11 +435 val_435 2010-03-03 11 +10 val_10 2010-03-03 11 +277 val_277 2010-03-03 11 +273 val_273 2010-03-03 11 +306 val_306 2010-03-03 11 +224 val_224 2010-03-03 11 +309 val_309 2010-03-03 11 +389 val_389 2010-03-03 11 +327 val_327 2010-03-03 11 +242 val_242 2010-03-03 11 +369 val_369 2010-03-03 11 +392 val_392 2010-03-03 11 272 val_272 2010-03-03 11 -373 val_373 2010-03-03 11 -353 val_353 2010-03-03 11 +331 val_331 2010-03-03 11 +401 val_401 2010-03-03 11 +242 val_242 2010-03-03 11 +452 val_452 2010-03-03 11 +177 val_177 2010-03-03 11 +226 val_226 2010-03-03 11 +5 val_5 2010-03-03 11 +497 val_497 2010-03-03 11 +402 val_402 2010-03-03 11 +396 val_396 2010-03-03 11 +317 val_317 2010-03-03 11 +395 val_395 2010-03-03 11 +58 val_58 2010-03-03 11 +35 val_35 2010-03-03 11 +336 val_336 2010-03-03 11 +95 val_95 2010-03-03 11 +11 val_11 2010-03-03 11 +168 val_168 2010-03-03 11 +34 val_34 2010-03-03 11 +229 val_229 2010-03-03 11 +233 val_233 2010-03-03 11 +143 val_143 2010-03-03 11 +472 val_472 2010-03-03 11 +322 val_322 2010-03-03 11 +498 val_498 2010-03-03 11 +160 val_160 2010-03-03 11 +195 val_195 2010-03-03 11 +42 val_42 2010-03-03 11 +321 val_321 2010-03-03 11 +430 val_430 2010-03-03 11 +119 val_119 2010-03-03 11 489 val_489 2010-03-03 11 -384 val_384 2010-03-03 11 -404 val_404 2010-03-03 11 -260 val_260 2010-03-03 11 -67 val_67 2010-03-03 11 -181 val_181 2010-03-03 11 +458 val_458 2010-03-03 11 +78 val_78 2010-03-03 11 +76 val_76 2010-03-03 11 +41 val_41 2010-03-03 11 +223 val_223 2010-03-03 11 +492 val_492 2010-03-03 11 +149 val_149 2010-03-03 11 +449 val_449 2010-03-03 11 +218 val_218 2010-03-03 11 +228 val_228 2010-03-03 11 +138 val_138 2010-03-03 11 +453 val_453 2010-03-03 11 +30 val_30 2010-03-03 11 +209 val_209 2010-03-03 11 +64 val_64 2010-03-03 11 +468 val_468 2010-03-03 11 +76 val_76 2010-03-03 11 +74 val_74 2010-03-03 11 +342 val_342 2010-03-03 11 +69 val_69 2010-03-03 11 230 val_230 2010-03-03 11 -12 val_12 2010-03-03 11 -284 val_284 2010-03-03 11 -180 val_180 2010-03-03 11 -333 val_333 2010-03-03 11 -83 val_83 2010-03-03 11 -241 val_241 2010-03-03 11 -137 val_137 2010-03-03 11 -275 val_275 2010-03-03 11 -311 val_311 2010-03-03 11 -332 val_332 2010-03-03 11 -318 val_318 2010-03-03 11 -65 val_65 2010-03-03 11 -221 val_221 2010-03-03 11 -289 val_289 2010-03-03 11 -176 val_176 2010-03-03 11 -278 val_278 2010-03-03 11 -430 val_430 2010-03-03 11 +33 val_33 2010-03-03 11 +368 val_368 2010-03-03 11 +103 val_103 2010-03-03 11 +296 val_296 2010-03-03 11 +113 val_113 2010-03-03 11 216 val_216 2010-03-03 11 -213 val_213 2010-03-03 11 +367 val_367 2010-03-03 11 +344 val_344 2010-03-03 11 +167 val_167 2010-03-03 11 +274 val_274 2010-03-03 11 +219 val_219 2010-03-03 11 239 val_239 2010-03-03 11 -103 val_103 2010-03-03 11 -138 val_138 2010-03-03 11 -51 val_51 2010-03-03 11 -459 val_459 2010-03-03 11 -54 val_54 2010-03-03 11 -176 val_176 2010-03-03 11 -187 val_187 2010-03-03 11 -286 val_286 2010-03-03 11 -192 val_192 2010-03-03 11 +485 val_485 2010-03-03 11 +116 val_116 2010-03-03 11 +223 val_223 2010-03-03 11 +256 val_256 2010-03-03 11 +263 val_263 2010-03-03 11 +70 val_70 2010-03-03 11 +487 val_487 2010-03-03 11 +480 val_480 2010-03-03 11 +401 val_401 2010-03-03 11 +288 val_288 2010-03-03 11 +191 val_191 2010-03-03 11 +5 val_5 2010-03-03 11 +244 val_244 2010-03-03 11 +438 val_438 2010-03-03 11 +128 val_128 2010-03-03 11 +467 val_467 2010-03-03 11 +432 val_432 2010-03-03 11 +202 val_202 2010-03-03 11 +316 val_316 2010-03-03 11 +229 val_229 2010-03-03 11 469 val_469 2010-03-03 11 -437 val_437 2010-03-03 11 -386 val_386 2010-03-03 11 +463 val_463 2010-03-03 11 +280 val_280 2010-03-03 11 +2 val_2 2010-03-03 11 +35 val_35 2010-03-03 11 +283 val_283 2010-03-03 11 +331 val_331 2010-03-03 11 +235 val_235 2010-03-03 11 +80 val_80 2010-03-03 11 +44 val_44 2010-03-03 11 +193 val_193 2010-03-03 11 +321 val_321 2010-03-03 11 +335 val_335 2010-03-03 11 +104 val_104 2010-03-03 11 +466 val_466 2010-03-03 11 +366 val_366 2010-03-03 11 +175 val_175 2010-03-03 11 +403 val_403 2010-03-03 11 +483 val_483 2010-03-03 11 +53 val_53 2010-03-03 11 +105 val_105 2010-03-03 11 +257 val_257 2010-03-03 11 +406 val_406 2010-03-03 11 +409 val_409 2010-03-03 11 +190 val_190 2010-03-03 11 +406 val_406 2010-03-03 11 +401 val_401 2010-03-03 11 +114 val_114 2010-03-03 11 +258 val_258 2010-03-03 11 +90 val_90 2010-03-03 11 +203 val_203 2010-03-03 11 +262 val_262 2010-03-03 11 +348 val_348 2010-03-03 11 +424 val_424 2010-03-03 11 +12 val_12 2010-03-03 11 +396 val_396 2010-03-03 11 +201 val_201 2010-03-03 11 +217 val_217 2010-03-03 11 +164 val_164 2010-03-03 11 +431 val_431 2010-03-03 11 +454 val_454 2010-03-03 11 +478 val_478 2010-03-03 11 +298 val_298 2010-03-03 11 125 val_125 2010-03-03 11 -498 val_498 2010-03-03 11 +431 val_431 2010-03-03 11 +164 val_164 2010-03-03 11 +424 val_424 2010-03-03 11 +187 val_187 2010-03-03 11 382 val_382 2010-03-03 11 -169 val_169 2010-03-03 11 -399 val_399 2010-03-03 11 -356 val_356 2010-03-03 11 -208 val_208 2010-03-03 11 -277 val_277 2010-03-03 11 -427 val_427 2010-03-03 11 -35 val_35 2010-03-03 11 -280 val_280 2010-03-03 11 -4 val_4 2010-03-03 11 -72 val_72 2010-03-03 11 -47 val_47 2010-03-03 11 -111 val_111 2010-03-03 11 -92 val_92 2010-03-03 11 -221 val_221 2010-03-03 11 -378 val_378 2010-03-03 11 -157 val_157 2010-03-03 11 -489 val_489 2010-03-03 11 -20 val_20 2010-03-03 11 -170 val_170 2010-03-03 11 -129 val_129 2010-03-03 11 -345 val_345 2010-03-03 11 +5 val_5 2010-03-03 11 +70 val_70 2010-03-03 11 +397 val_397 2010-03-03 11 +480 val_480 2010-03-03 11 +291 val_291 2010-03-03 11 +24 val_24 2010-03-03 11 +351 val_351 2010-03-03 11 +255 val_255 2010-03-03 11 +104 val_104 2010-03-03 11 +70 val_70 2010-03-03 11 +163 val_163 2010-03-03 11 438 val_438 2010-03-03 11 -149 val_149 2010-03-03 11 -205 val_205 2010-03-03 11 -302 val_302 2010-03-03 11 -57 val_57 2010-03-03 11 -316 val_316 2010-03-03 11 -311 val_311 2010-03-03 11 -128 val_128 2010-03-03 11 -455 val_455 2010-03-03 11 -0 val_0 2010-03-03 11 -339 val_339 2010-03-03 11 -203 val_203 2010-03-03 11 -155 val_155 2010-03-03 11 -113 val_113 2010-03-03 11 -17 val_17 2010-03-03 11 -475 val_475 2010-03-03 11 -195 val_195 2010-03-03 11 -167 val_167 2010-03-03 11 -325 val_325 2010-03-03 11 -367 val_367 2010-03-03 11 -342 val_342 2010-03-03 11 +119 val_119 2010-03-03 11 +414 val_414 2010-03-03 11 +200 val_200 2010-03-03 11 +491 val_491 2010-03-03 11 +237 val_237 2010-03-03 11 439 val_439 2010-03-03 11 -266 val_266 2010-03-03 11 -365 val_365 2010-03-03 11 -309 val_309 2010-03-03 11 -397 val_397 2010-03-03 11 -377 val_377 2010-03-03 11 -162 val_162 2010-03-03 11 -489 val_489 2010-03-03 11 +360 val_360 2010-03-03 11 +248 val_248 2010-03-03 11 +479 val_479 2010-03-03 11 +305 val_305 2010-03-03 11 417 val_417 2010-03-03 11 -247 val_247 2010-03-03 11 -396 val_396 2010-03-03 11 -399 val_399 2010-03-03 11 -174 val_174 2010-03-03 11 -208 val_208 2010-03-03 11 -466 val_466 2010-03-03 11 199 val_199 2010-03-03 11 -207 val_207 2010-03-03 11 -494 val_494 2010-03-03 11 -413 val_413 2010-03-03 11 -174 val_174 2010-03-03 11 -482 val_482 2010-03-03 11 -237 val_237 2010-03-03 11 -394 val_394 2010-03-03 11 -459 val_459 2010-03-03 11 -446 val_446 2010-03-03 11 -338 val_338 2010-03-03 11 -193 val_193 2010-03-03 11 -153 val_153 2010-03-03 11 -287 val_287 2010-03-03 11 -219 val_219 2010-03-03 11 -292 val_292 2010-03-03 11 -252 val_252 2010-03-03 11 -430 val_430 2010-03-03 11 -417 val_417 2010-03-03 11 -166 val_166 2010-03-03 11 -403 val_403 2010-03-03 11 -82 val_82 2010-03-03 11 -15 val_15 2010-03-03 11 -209 val_209 2010-03-03 11 -277 val_277 2010-03-03 11 -281 val_281 2010-03-03 11 -327 val_327 2010-03-03 11 -37 val_37 2010-03-03 11 -495 val_495 2010-03-03 11 -145 val_145 2010-03-03 11 -469 val_469 2010-03-03 11 -152 val_152 2010-03-03 11 -374 val_374 2010-03-03 11 +444 val_444 2010-03-03 11 +120 val_120 2010-03-03 11 429 val_429 2010-03-03 11 +169 val_169 2010-03-03 11 +443 val_443 2010-03-03 11 +323 val_323 2010-03-03 11 +325 val_325 2010-03-03 11 +277 val_277 2010-03-03 11 +230 val_230 2010-03-03 11 +478 val_478 2010-03-03 11 +178 val_178 2010-03-03 11 +468 val_468 2010-03-03 11 +310 val_310 2010-03-03 11 +317 val_317 2010-03-03 11 +333 val_333 2010-03-03 11 +493 val_493 2010-03-03 11 +460 val_460 2010-03-03 11 +207 val_207 2010-03-03 11 +249 val_249 2010-03-03 11 +265 val_265 2010-03-03 11 +480 val_480 2010-03-03 11 +83 val_83 2010-03-03 11 +136 val_136 2010-03-03 11 +353 val_353 2010-03-03 11 +172 val_172 2010-03-03 11 +214 val_214 2010-03-03 11 +462 val_462 2010-03-03 11 +233 val_233 2010-03-03 11 406 val_406 2010-03-03 11 +133 val_133 2010-03-03 11 +175 val_175 2010-03-03 11 +189 val_189 2010-03-03 11 +454 val_454 2010-03-03 11 +375 val_375 2010-03-03 11 +401 val_401 2010-03-03 11 +421 val_421 2010-03-03 11 +407 val_407 2010-03-03 11 +384 val_384 2010-03-03 11 +256 val_256 2010-03-03 11 +26 val_26 2010-03-03 11 +134 val_134 2010-03-03 11 +67 val_67 2010-03-03 11 +384 val_384 2010-03-03 11 +379 val_379 2010-03-03 11 +18 val_18 2010-03-03 11 +462 val_462 2010-03-03 11 +492 val_492 2010-03-03 11 +100 val_100 2010-03-03 11 +298 val_298 2010-03-03 11 +9 val_9 2010-03-03 11 +341 val_341 2010-03-03 11 +498 val_498 2010-03-03 11 146 val_146 2010-03-03 11 -213 val_213 2010-03-03 11 -128 val_128 2010-03-03 11 -66 val_66 2010-03-03 11 -369 val_369 2010-03-03 11 -224 val_224 2010-03-03 11 +458 val_458 2010-03-03 11 +362 val_362 2010-03-03 11 +186 val_186 2010-03-03 11 +285 val_285 2010-03-03 11 +348 val_348 2010-03-03 11 +167 val_167 2010-03-03 11 +18 val_18 2010-03-03 11 273 val_273 2010-03-03 11 -150 val_150 2010-03-03 11 -401 val_401 2010-03-03 11 -193 val_193 2010-03-03 11 -265 val_265 2010-03-03 11 -484 val_484 2010-03-03 11 -98 val_98 2010-03-03 11 -278 val_278 2010-03-03 11 -255 val_255 2010-03-03 11 -409 val_409 2010-03-03 11 -165 val_165 2010-03-03 11 -27 val_27 2010-03-03 11 -311 val_311 2010-03-03 11 -86 val_86 2010-03-03 11 +183 val_183 2010-03-03 11 +281 val_281 2010-03-03 11 +344 val_344 2010-03-03 11 97 val_97 2010-03-03 11 -200 val_200 2010-03-03 11 -400 val_400 2010-03-03 11 -403 val_403 2010-03-03 11 -169 val_169 2010-03-03 11 -90 val_90 2010-03-03 11 -126 val_126 2010-03-03 11 -222 val_222 2010-03-03 11 -477 val_477 2010-03-03 11 -414 val_414 2010-03-03 11 -194 val_194 2010-03-03 11 -307 val_307 2010-03-03 11 -348 val_348 2010-03-03 11 -152 val_152 2010-03-03 11 -448 val_448 2010-03-03 11 -37 val_37 2010-03-03 11 -28 val_28 2010-03-03 11 -84 val_84 2010-03-03 11 -315 val_315 2010-03-03 11 469 val_469 2010-03-03 11 +315 val_315 2010-03-03 11 +84 val_84 2010-03-03 11 +28 val_28 2010-03-03 11 +37 val_37 2010-03-03 11 +448 val_448 2010-03-03 11 +152 val_152 2010-03-03 11 +348 val_348 2010-03-03 11 +307 val_307 2010-03-03 11 +194 val_194 2010-03-03 11 +414 val_414 2010-03-03 11 +477 val_477 2010-03-03 11 +222 val_222 2010-03-03 11 +126 val_126 2010-03-03 11 +90 val_90 2010-03-03 11 +169 val_169 2010-03-03 11 +403 val_403 2010-03-03 11 +400 val_400 2010-03-03 11 +200 val_200 2010-03-03 11 97 val_97 2010-03-03 11 -344 val_344 2010-03-03 11 -281 val_281 2010-03-03 11 -183 val_183 2010-03-03 11 +238 val_238 2010-03-03 11 +86 val_86 2010-03-03 11 +311 val_311 2010-03-03 11 +27 val_27 2010-03-03 11 +165 val_165 2010-03-03 11 +409 val_409 2010-03-03 11 +255 val_255 2010-03-03 11 +278 val_278 2010-03-03 11 +98 val_98 2010-03-03 11 +484 val_484 2010-03-03 11 +265 val_265 2010-03-03 11 +193 val_193 2010-03-03 11 +401 val_401 2010-03-03 11 +150 val_150 2010-03-03 11 273 val_273 2010-03-03 11 -18 val_18 2010-03-03 11 -167 val_167 2010-03-03 11 -348 val_348 2010-03-03 11 -285 val_285 2010-03-03 11 -186 val_186 2010-03-03 11 -362 val_362 2010-03-03 11 -458 val_458 2010-03-03 11 +224 val_224 2010-03-03 11 +369 val_369 2010-03-03 11 +66 val_66 2010-03-03 11 +128 val_128 2010-03-03 11 +213 val_213 2010-03-03 11 146 val_146 2010-03-03 11 -498 val_498 2010-03-03 11 -341 val_341 2010-03-03 11 -9 val_9 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -492 val_492 2010-03-03 11 -462 val_462 2010-03-03 11 -18 val_18 2010-03-03 11 -379 val_379 2010-03-03 11 -384 val_384 2010-03-03 11 -67 val_67 2010-03-03 11 -134 val_134 2010-03-03 11 -26 val_26 2010-03-03 11 -256 val_256 2010-03-03 11 -384 val_384 2010-03-03 11 -407 val_407 2010-03-03 11 -421 val_421 2010-03-03 11 -401 val_401 2010-03-03 11 -375 val_375 2010-03-03 11 -454 val_454 2010-03-03 11 -189 val_189 2010-03-03 11 -175 val_175 2010-03-03 11 -133 val_133 2010-03-03 11 406 val_406 2010-03-03 11 -233 val_233 2010-03-03 11 -462 val_462 2010-03-03 11 -214 val_214 2010-03-03 11 -172 val_172 2010-03-03 11 -353 val_353 2010-03-03 11 -136 val_136 2010-03-03 11 -83 val_83 2010-03-03 11 -480 val_480 2010-03-03 11 -265 val_265 2010-03-03 11 -249 val_249 2010-03-03 11 -207 val_207 2010-03-03 11 -460 val_460 2010-03-03 11 -493 val_493 2010-03-03 11 -333 val_333 2010-03-03 11 -317 val_317 2010-03-03 11 -310 val_310 2010-03-03 11 -468 val_468 2010-03-03 11 -178 val_178 2010-03-03 11 -478 val_478 2010-03-03 11 -230 val_230 2010-03-03 11 -277 val_277 2010-03-03 11 -325 val_325 2010-03-03 11 -323 val_323 2010-03-03 11 -443 val_443 2010-03-03 11 -169 val_169 2010-03-03 11 429 val_429 2010-03-03 11 -120 val_120 2010-03-03 11 -444 val_444 2010-03-03 11 -199 val_199 2010-03-03 11 +374 val_374 2010-03-03 11 +152 val_152 2010-03-03 11 +469 val_469 2010-03-03 11 +145 val_145 2010-03-03 11 +495 val_495 2010-03-03 11 +37 val_37 2010-03-03 11 +327 val_327 2010-03-03 11 +281 val_281 2010-03-03 11 +277 val_277 2010-03-03 11 +209 val_209 2010-03-03 11 +15 val_15 2010-03-03 11 +82 val_82 2010-03-03 11 +403 val_403 2010-03-03 11 +166 val_166 2010-03-03 11 417 val_417 2010-03-03 11 -305 val_305 2010-03-03 11 -479 val_479 2010-03-03 11 -248 val_248 2010-03-03 11 -360 val_360 2010-03-03 11 -439 val_439 2010-03-03 11 +430 val_430 2010-03-03 11 +252 val_252 2010-03-03 11 +292 val_292 2010-03-03 11 +219 val_219 2010-03-03 11 +287 val_287 2010-03-03 11 +153 val_153 2010-03-03 11 +193 val_193 2010-03-03 11 +338 val_338 2010-03-03 11 +446 val_446 2010-03-03 11 +459 val_459 2010-03-03 11 +394 val_394 2010-03-03 11 237 val_237 2010-03-03 11 -491 val_491 2010-03-03 11 -200 val_200 2010-03-03 11 -414 val_414 2010-03-03 11 -119 val_119 2010-03-03 11 -438 val_438 2010-03-03 11 -163 val_163 2010-03-03 11 -70 val_70 2010-03-03 11 -104 val_104 2010-03-03 11 -255 val_255 2010-03-03 11 -351 val_351 2010-03-03 11 -24 val_24 2010-03-03 11 -291 val_291 2010-03-03 11 -480 val_480 2010-03-03 11 -397 val_397 2010-03-03 11 -70 val_70 2010-03-03 11 -5 val_5 2010-03-03 11 -382 val_382 2010-03-03 11 -187 val_187 2010-03-03 11 -424 val_424 2010-03-03 11 -164 val_164 2010-03-03 11 -431 val_431 2010-03-03 11 -125 val_125 2010-03-03 11 -298 val_298 2010-03-03 11 -478 val_478 2010-03-03 11 -454 val_454 2010-03-03 11 -431 val_431 2010-03-03 11 -164 val_164 2010-03-03 11 -217 val_217 2010-03-03 11 -201 val_201 2010-03-03 11 +482 val_482 2010-03-03 11 +174 val_174 2010-03-03 11 +413 val_413 2010-03-03 11 +494 val_494 2010-03-03 11 +207 val_207 2010-03-03 11 +199 val_199 2010-03-03 11 +466 val_466 2010-03-03 11 +208 val_208 2010-03-03 11 +174 val_174 2010-03-03 11 +399 val_399 2010-03-03 11 396 val_396 2010-03-03 11 -12 val_12 2010-03-03 11 -424 val_424 2010-03-03 11 -348 val_348 2010-03-03 11 -262 val_262 2010-03-03 11 +247 val_247 2010-03-03 11 +417 val_417 2010-03-03 11 +489 val_489 2010-03-03 11 +162 val_162 2010-03-03 11 +377 val_377 2010-03-03 11 +397 val_397 2010-03-03 11 +309 val_309 2010-03-03 11 +365 val_365 2010-03-03 11 +266 val_266 2010-03-03 11 +439 val_439 2010-03-03 11 +342 val_342 2010-03-03 11 +367 val_367 2010-03-03 11 +325 val_325 2010-03-03 11 +167 val_167 2010-03-03 11 +195 val_195 2010-03-03 11 +475 val_475 2010-03-03 11 +17 val_17 2010-03-03 11 +113 val_113 2010-03-03 11 +155 val_155 2010-03-03 11 203 val_203 2010-03-03 11 -90 val_90 2010-03-03 11 -258 val_258 2010-03-03 11 -114 val_114 2010-03-03 11 -401 val_401 2010-03-03 11 -406 val_406 2010-03-03 11 -190 val_190 2010-03-03 11 -409 val_409 2010-03-03 11 -406 val_406 2010-03-03 11 -257 val_257 2010-03-03 11 -105 val_105 2010-03-03 11 -53 val_53 2010-03-03 11 -483 val_483 2010-03-03 11 -403 val_403 2010-03-03 11 -175 val_175 2010-03-03 11 -366 val_366 2010-03-03 11 -466 val_466 2010-03-03 11 -104 val_104 2010-03-03 11 -335 val_335 2010-03-03 11 -321 val_321 2010-03-03 11 -193 val_193 2010-03-03 11 -44 val_44 2010-03-03 11 -80 val_80 2010-03-03 11 -235 val_235 2010-03-03 11 -331 val_331 2010-03-03 11 -283 val_283 2010-03-03 11 -35 val_35 2010-03-03 11 -2 val_2 2010-03-03 11 -280 val_280 2010-03-03 11 -463 val_463 2010-03-03 11 -469 val_469 2010-03-03 11 -229 val_229 2010-03-03 11 -316 val_316 2010-03-03 11 -202 val_202 2010-03-03 11 -432 val_432 2010-03-03 11 -467 val_467 2010-03-03 11 +339 val_339 2010-03-03 11 +0 val_0 2010-03-03 11 +455 val_455 2010-03-03 11 128 val_128 2010-03-03 11 +311 val_311 2010-03-03 11 +316 val_316 2010-03-03 11 +57 val_57 2010-03-03 11 +302 val_302 2010-03-03 11 +205 val_205 2010-03-03 11 +149 val_149 2010-03-03 11 438 val_438 2010-03-03 11 -244 val_244 2010-03-03 11 -5 val_5 2010-03-03 11 -191 val_191 2010-03-03 11 -288 val_288 2010-03-03 11 -401 val_401 2010-03-03 11 -480 val_480 2010-03-03 11 -487 val_487 2010-03-03 11 -70 val_70 2010-03-03 11 -263 val_263 2010-03-03 11 -256 val_256 2010-03-03 11 -223 val_223 2010-03-03 11 -116 val_116 2010-03-03 11 -485 val_485 2010-03-03 11 +345 val_345 2010-03-03 11 +129 val_129 2010-03-03 11 +170 val_170 2010-03-03 11 +20 val_20 2010-03-03 11 +489 val_489 2010-03-03 11 +157 val_157 2010-03-03 11 +378 val_378 2010-03-03 11 +221 val_221 2010-03-03 11 +92 val_92 2010-03-03 11 +111 val_111 2010-03-03 11 +47 val_47 2010-03-03 11 +72 val_72 2010-03-03 11 +4 val_4 2010-03-03 11 +280 val_280 2010-03-03 11 +35 val_35 2010-03-03 11 +427 val_427 2010-03-03 11 +277 val_277 2010-03-03 11 +208 val_208 2010-03-03 11 +356 val_356 2010-03-03 11 +399 val_399 2010-03-03 11 +169 val_169 2010-03-03 11 +382 val_382 2010-03-03 11 +498 val_498 2010-03-03 11 +125 val_125 2010-03-03 11 +386 val_386 2010-03-03 11 +437 val_437 2010-03-03 11 +469 val_469 2010-03-03 11 +192 val_192 2010-03-03 11 +286 val_286 2010-03-03 11 +187 val_187 2010-03-03 11 +176 val_176 2010-03-03 11 +54 val_54 2010-03-03 11 +459 val_459 2010-03-03 11 +51 val_51 2010-03-03 11 +138 val_138 2010-03-03 11 +103 val_103 2010-03-03 11 239 val_239 2010-03-03 11 -219 val_219 2010-03-03 11 -274 val_274 2010-03-03 11 -167 val_167 2010-03-03 11 -344 val_344 2010-03-03 11 -367 val_367 2010-03-03 11 +213 val_213 2010-03-03 11 216 val_216 2010-03-03 11 -113 val_113 2010-03-03 11 -296 val_296 2010-03-03 11 -103 val_103 2010-03-03 11 -368 val_368 2010-03-03 11 -33 val_33 2010-03-03 11 +430 val_430 2010-03-03 11 +278 val_278 2010-03-03 11 +176 val_176 2010-03-03 11 +289 val_289 2010-03-03 11 +221 val_221 2010-03-03 11 +65 val_65 2010-03-03 11 +318 val_318 2010-03-03 11 +332 val_332 2010-03-03 11 +311 val_311 2010-03-03 11 +275 val_275 2010-03-03 11 +137 val_137 2010-03-03 11 +241 val_241 2010-03-03 11 +83 val_83 2010-03-03 11 +333 val_333 2010-03-03 11 +180 val_180 2010-03-03 11 +284 val_284 2010-03-03 11 +12 val_12 2010-03-03 11 230 val_230 2010-03-03 11 -69 val_69 2010-03-03 11 -342 val_342 2010-03-03 11 -74 val_74 2010-03-03 11 -76 val_76 2010-03-03 11 -468 val_468 2010-03-03 11 -64 val_64 2010-03-03 11 -209 val_209 2010-03-03 11 -30 val_30 2010-03-03 11 -453 val_453 2010-03-03 11 -138 val_138 2010-03-03 11 -228 val_228 2010-03-03 11 -218 val_218 2010-03-03 11 -449 val_449 2010-03-03 11 -149 val_149 2010-03-03 11 -492 val_492 2010-03-03 11 -223 val_223 2010-03-03 11 -41 val_41 2010-03-03 11 -76 val_76 2010-03-03 11 -78 val_78 2010-03-03 11 -458 val_458 2010-03-03 11 +181 val_181 2010-03-03 11 +67 val_67 2010-03-03 11 +260 val_260 2010-03-03 11 +404 val_404 2010-03-03 11 +384 val_384 2010-03-03 11 489 val_489 2010-03-03 11 -119 val_119 2010-03-03 11 -430 val_430 2010-03-03 11 -321 val_321 2010-03-03 11 +353 val_353 2010-03-03 11 +373 val_373 2010-03-03 11 +272 val_272 2010-03-03 11 +138 val_138 2010-03-03 11 +217 val_217 2010-03-03 11 +84 val_84 2010-03-03 11 +348 val_348 2010-03-03 11 +466 val_466 2010-03-03 11 +58 val_58 2010-03-03 11 +8 val_8 2010-03-03 11 +411 val_411 2010-03-03 11 +230 val_230 2010-03-03 11 +208 val_208 2010-03-03 11 +348 val_348 2010-03-03 11 +24 val_24 2010-03-03 11 +463 val_463 2010-03-03 11 +431 val_431 2010-03-03 11 +179 val_179 2010-03-03 11 +172 val_172 2010-03-03 11 42 val_42 2010-03-03 11 -195 val_195 2010-03-03 11 -160 val_160 2010-03-03 11 -498 val_498 2010-03-03 11 +129 val_129 2010-03-03 11 +158 val_158 2010-03-03 11 +119 val_119 2010-03-03 11 +496 val_496 2010-03-03 11 +0 val_0 2010-03-03 11 322 val_322 2010-03-03 11 -472 val_472 2010-03-03 11 -143 val_143 2010-03-03 11 -233 val_233 2010-03-03 11 -229 val_229 2010-03-03 11 -34 val_34 2010-03-03 11 -168 val_168 2010-03-03 11 -11 val_11 2010-03-03 11 -95 val_95 2010-03-03 11 -336 val_336 2010-03-03 11 -35 val_35 2010-03-03 11 -58 val_58 2010-03-03 11 -395 val_395 2010-03-03 11 -317 val_317 2010-03-03 11 -396 val_396 2010-03-03 11 -402 val_402 2010-03-03 11 -497 val_497 2010-03-03 11 -5 val_5 2010-03-03 11 -226 val_226 2010-03-03 11 -177 val_177 2010-03-03 11 -452 val_452 2010-03-03 11 -242 val_242 2010-03-03 11 -401 val_401 2010-03-03 11 -331 val_331 2010-03-03 11 -272 val_272 2010-03-03 11 -392 val_392 2010-03-03 11 -369 val_369 2010-03-03 11 -242 val_242 2010-03-03 11 -327 val_327 2010-03-03 11 -389 val_389 2010-03-03 11 -309 val_309 2010-03-03 11 -224 val_224 2010-03-03 11 -306 val_306 2010-03-03 11 -273 val_273 2010-03-03 11 -277 val_277 2010-03-03 11 -10 val_10 2010-03-03 11 -435 val_435 2010-03-03 11 -19 val_19 2010-03-03 11 -307 val_307 2010-03-03 11 -90 val_90 2010-03-03 11 -72 val_72 2010-03-03 11 -118 val_118 2010-03-03 11 -15 val_15 2010-03-03 11 -419 val_419 2010-03-03 11 -238 val_238 2010-03-03 11 -138 val_138 2010-03-03 11 -282 val_282 2010-03-03 11 -395 val_395 2010-03-03 11 -134 val_134 2010-03-03 11 -118 val_118 2010-03-03 11 -179 val_179 2010-03-03 11 -364 val_364 2010-03-03 11 -87 val_87 2010-03-03 11 -490 val_490 2010-03-03 11 -0 val_0 2010-03-03 11 -77 val_77 2010-03-03 11 -85 val_85 2010-03-03 11 -413 val_413 2010-03-03 11 -169 val_169 2010-03-03 11 -316 val_316 2010-03-03 11 -369 val_369 2010-03-03 11 -137 val_137 2010-03-03 11 -470 val_470 2010-03-03 11 -409 val_409 2010-03-03 11 -318 val_318 2010-03-03 11 -318 val_318 2010-03-03 11 -187 val_187 2010-03-03 11 197 val_197 2010-03-03 11 -282 val_282 2010-03-03 11 -98 val_98 2010-03-03 11 -457 val_457 2010-03-03 11 -481 val_481 2010-03-03 11 -288 val_288 2010-03-03 11 -196 val_196 2010-03-03 11 -95 val_95 2010-03-03 11 -308 val_308 2010-03-03 11 468 val_468 2010-03-03 11 -469 val_469 2010-03-03 11 -156 val_156 2010-03-03 11 -436 val_436 2010-03-03 11 -43 val_43 2010-03-03 11 -404 val_404 2010-03-03 11 -51 val_51 2010-03-03 11 -131 val_131 2010-03-03 11 -120 val_120 2010-03-03 11 -205 val_205 2010-03-03 11 -230 val_230 2010-03-03 11 -327 val_327 2010-03-03 11 -165 val_165 2010-03-03 11 -26 val_26 2010-03-03 11 -96 val_96 2010-03-03 11 -418 val_418 2010-03-03 11 -191 val_191 2010-03-03 11 -199 val_199 2010-03-03 11 -298 val_298 2010-03-03 11 -100 val_100 2010-03-03 11 -454 val_454 2010-03-03 11 393 val_393 2010-03-03 11 +454 val_454 2010-03-03 11 +100 val_100 2010-03-03 11 +298 val_298 2010-03-03 11 +199 val_199 2010-03-03 11 +191 val_191 2010-03-03 11 +418 val_418 2010-03-03 11 +96 val_96 2010-03-03 11 +26 val_26 2010-03-03 11 +165 val_165 2010-03-03 11 +327 val_327 2010-03-03 11 +230 val_230 2010-03-03 11 +205 val_205 2010-03-03 11 +120 val_120 2010-03-03 11 +131 val_131 2010-03-03 11 +51 val_51 2010-03-03 11 +404 val_404 2010-03-03 11 +43 val_43 2010-03-03 11 +436 val_436 2010-03-03 11 +156 val_156 2010-03-03 11 +469 val_469 2010-03-03 11 468 val_468 2010-03-03 11 +308 val_308 2010-03-03 11 +95 val_95 2010-03-03 11 +196 val_196 2010-03-03 11 +288 val_288 2010-03-03 11 +481 val_481 2010-03-03 11 +457 val_457 2010-03-03 11 +98 val_98 2010-03-03 11 +282 val_282 2010-03-03 11 197 val_197 2010-03-03 11 -322 val_322 2010-03-03 11 +187 val_187 2010-03-03 11 +318 val_318 2010-03-03 11 +318 val_318 2010-03-03 11 +409 val_409 2010-03-03 11 +470 val_470 2010-03-03 11 +137 val_137 2010-03-03 11 +369 val_369 2010-03-03 11 +316 val_316 2010-03-03 11 +169 val_169 2010-03-03 11 +413 val_413 2010-03-03 11 +85 val_85 2010-03-03 11 +77 val_77 2010-03-03 11 0 val_0 2010-03-03 11 -496 val_496 2010-03-03 11 -119 val_119 2010-03-03 11 -158 val_158 2010-03-03 11 -129 val_129 2010-03-03 11 -42 val_42 2010-03-03 11 -172 val_172 2010-03-03 11 +490 val_490 2010-03-03 11 +87 val_87 2010-03-03 11 +364 val_364 2010-03-03 11 179 val_179 2010-03-03 11 -431 val_431 2010-03-03 11 -463 val_463 2010-03-03 11 -24 val_24 2010-03-03 11 -348 val_348 2010-03-03 11 -208 val_208 2010-03-03 11 -230 val_230 2010-03-03 11 -411 val_411 2010-03-03 11 -8 val_8 2010-03-03 11 -58 val_58 2010-03-03 11 -466 val_466 2010-03-03 11 -348 val_348 2010-03-03 11 -84 val_84 2010-03-03 11 -217 val_217 2010-03-03 11 +118 val_118 2010-03-03 11 +134 val_134 2010-03-03 11 +395 val_395 2010-03-03 11 +282 val_282 2010-03-03 11 138 val_138 2010-03-03 11 +238 val_238 2010-03-03 11 +419 val_419 2010-03-03 11 +15 val_15 2010-03-03 11 +118 val_118 2010-03-03 11 +72 val_72 2010-03-03 11 +90 val_90 2010-03-03 11 +307 val_307 2010-03-03 11 +19 val_19 2010-03-03 11 +435 val_435 2010-03-03 11 +10 val_10 2010-03-03 11 +277 val_277 2010-03-03 11 +273 val_273 2010-03-03 11 +306 val_306 2010-03-03 11 +224 val_224 2010-03-03 11 +309 val_309 2010-03-03 11 +389 val_389 2010-03-03 11 +327 val_327 2010-03-03 11 +242 val_242 2010-03-03 11 +369 val_369 2010-03-03 11 +392 val_392 2010-03-03 11 272 val_272 2010-03-03 11 -373 val_373 2010-03-03 11 -353 val_353 2010-03-03 11 +331 val_331 2010-03-03 11 +401 val_401 2010-03-03 11 +242 val_242 2010-03-03 11 +452 val_452 2010-03-03 11 +177 val_177 2010-03-03 11 +226 val_226 2010-03-03 11 +5 val_5 2010-03-03 11 +497 val_497 2010-03-03 11 +402 val_402 2010-03-03 11 +396 val_396 2010-03-03 11 +317 val_317 2010-03-03 11 +395 val_395 2010-03-03 11 +58 val_58 2010-03-03 11 +35 val_35 2010-03-03 11 +336 val_336 2010-03-03 11 +95 val_95 2010-03-03 11 +11 val_11 2010-03-03 11 +168 val_168 2010-03-03 11 +34 val_34 2010-03-03 11 +229 val_229 2010-03-03 11 +233 val_233 2010-03-03 11 +143 val_143 2010-03-03 11 +472 val_472 2010-03-03 11 +322 val_322 2010-03-03 11 +498 val_498 2010-03-03 11 +160 val_160 2010-03-03 11 +195 val_195 2010-03-03 11 +42 val_42 2010-03-03 11 +321 val_321 2010-03-03 11 +430 val_430 2010-03-03 11 +119 val_119 2010-03-03 11 489 val_489 2010-03-03 11 -384 val_384 2010-03-03 11 -404 val_404 2010-03-03 11 -260 val_260 2010-03-03 11 -67 val_67 2010-03-03 11 -181 val_181 2010-03-03 11 +458 val_458 2010-03-03 11 +78 val_78 2010-03-03 11 +76 val_76 2010-03-03 11 +41 val_41 2010-03-03 11 +223 val_223 2010-03-03 11 +492 val_492 2010-03-03 11 +149 val_149 2010-03-03 11 +449 val_449 2010-03-03 11 +218 val_218 2010-03-03 11 +228 val_228 2010-03-03 11 +138 val_138 2010-03-03 11 +453 val_453 2010-03-03 11 +30 val_30 2010-03-03 11 +209 val_209 2010-03-03 11 +64 val_64 2010-03-03 11 +468 val_468 2010-03-03 11 +76 val_76 2010-03-03 11 +74 val_74 2010-03-03 11 +342 val_342 2010-03-03 11 +69 val_69 2010-03-03 11 230 val_230 2010-03-03 11 -12 val_12 2010-03-03 11 -284 val_284 2010-03-03 11 -180 val_180 2010-03-03 11 -333 val_333 2010-03-03 11 -83 val_83 2010-03-03 11 -241 val_241 2010-03-03 11 -137 val_137 2010-03-03 11 -275 val_275 2010-03-03 11 -311 val_311 2010-03-03 11 -332 val_332 2010-03-03 11 -318 val_318 2010-03-03 11 -65 val_65 2010-03-03 11 -221 val_221 2010-03-03 11 -289 val_289 2010-03-03 11 -176 val_176 2010-03-03 11 -278 val_278 2010-03-03 11 -430 val_430 2010-03-03 11 -216 val_216 2010-03-03 11 -213 val_213 2010-03-03 11 -239 val_239 2010-03-03 11 +33 val_33 2010-03-03 11 +368 val_368 2010-03-03 11 103 val_103 2010-03-03 11 -138 val_138 2010-03-03 11 -51 val_51 2010-03-03 11 -459 val_459 2010-03-03 11 -54 val_54 2010-03-03 11 -176 val_176 2010-03-03 11 -187 val_187 2010-03-03 11 -286 val_286 2010-03-03 11 -192 val_192 2010-03-03 11 +296 val_296 2010-03-03 11 +113 val_113 2010-03-03 11 +216 val_216 2010-03-03 11 +367 val_367 2010-03-03 11 +344 val_344 2010-03-03 11 +167 val_167 2010-03-03 11 +274 val_274 2010-03-03 11 +219 val_219 2010-03-03 11 +239 val_239 2010-03-03 11 +485 val_485 2010-03-03 11 +116 val_116 2010-03-03 11 +223 val_223 2010-03-03 11 +256 val_256 2010-03-03 11 +263 val_263 2010-03-03 11 +70 val_70 2010-03-03 11 +487 val_487 2010-03-03 11 +480 val_480 2010-03-03 11 +401 val_401 2010-03-03 11 +288 val_288 2010-03-03 11 +191 val_191 2010-03-03 11 +5 val_5 2010-03-03 11 +244 val_244 2010-03-03 11 +438 val_438 2010-03-03 11 +128 val_128 2010-03-03 11 +467 val_467 2010-03-03 11 +432 val_432 2010-03-03 11 +202 val_202 2010-03-03 11 +316 val_316 2010-03-03 11 +229 val_229 2010-03-03 11 469 val_469 2010-03-03 11 -437 val_437 2010-03-03 11 -386 val_386 2010-03-03 11 +463 val_463 2010-03-03 11 +280 val_280 2010-03-03 11 +2 val_2 2010-03-03 11 +35 val_35 2010-03-03 11 +283 val_283 2010-03-03 11 +331 val_331 2010-03-03 11 +235 val_235 2010-03-03 11 +80 val_80 2010-03-03 11 +44 val_44 2010-03-03 11 +193 val_193 2010-03-03 11 +321 val_321 2010-03-03 11 +335 val_335 2010-03-03 11 +104 val_104 2010-03-03 11 +466 val_466 2010-03-03 11 +366 val_366 2010-03-03 11 +175 val_175 2010-03-03 11 +403 val_403 2010-03-03 11 +483 val_483 2010-03-03 11 +53 val_53 2010-03-03 11 +105 val_105 2010-03-03 11 +257 val_257 2010-03-03 11 +406 val_406 2010-03-03 11 +409 val_409 2010-03-03 11 +190 val_190 2010-03-03 11 +406 val_406 2010-03-03 11 +401 val_401 2010-03-03 11 +114 val_114 2010-03-03 11 +258 val_258 2010-03-03 11 +90 val_90 2010-03-03 11 +203 val_203 2010-03-03 11 +262 val_262 2010-03-03 11 +348 val_348 2010-03-03 11 +424 val_424 2010-03-03 11 +12 val_12 2010-03-03 11 +396 val_396 2010-03-03 11 +201 val_201 2010-03-03 11 +217 val_217 2010-03-03 11 +164 val_164 2010-03-03 11 +431 val_431 2010-03-03 11 +454 val_454 2010-03-03 11 +478 val_478 2010-03-03 11 +298 val_298 2010-03-03 11 125 val_125 2010-03-03 11 -498 val_498 2010-03-03 11 +431 val_431 2010-03-03 11 +164 val_164 2010-03-03 11 +424 val_424 2010-03-03 11 +187 val_187 2010-03-03 11 382 val_382 2010-03-03 11 -169 val_169 2010-03-03 11 -399 val_399 2010-03-03 11 -356 val_356 2010-03-03 11 -208 val_208 2010-03-03 11 -277 val_277 2010-03-03 11 -427 val_427 2010-03-03 11 -35 val_35 2010-03-03 11 -280 val_280 2010-03-03 11 -4 val_4 2010-03-03 11 -72 val_72 2010-03-03 11 -47 val_47 2010-03-03 11 -111 val_111 2010-03-03 11 -92 val_92 2010-03-03 11 -221 val_221 2010-03-03 11 -378 val_378 2010-03-03 11 -157 val_157 2010-03-03 11 -489 val_489 2010-03-03 11 -20 val_20 2010-03-03 11 -170 val_170 2010-03-03 11 -129 val_129 2010-03-03 11 -345 val_345 2010-03-03 11 +5 val_5 2010-03-03 11 +70 val_70 2010-03-03 11 +397 val_397 2010-03-03 11 +480 val_480 2010-03-03 11 +291 val_291 2010-03-03 11 +24 val_24 2010-03-03 11 +351 val_351 2010-03-03 11 +255 val_255 2010-03-03 11 +104 val_104 2010-03-03 11 +70 val_70 2010-03-03 11 +163 val_163 2010-03-03 11 438 val_438 2010-03-03 11 -149 val_149 2010-03-03 11 -205 val_205 2010-03-03 11 -302 val_302 2010-03-03 11 -57 val_57 2010-03-03 11 -316 val_316 2010-03-03 11 -311 val_311 2010-03-03 11 -128 val_128 2010-03-03 11 -455 val_455 2010-03-03 11 -0 val_0 2010-03-03 11 -339 val_339 2010-03-03 11 -203 val_203 2010-03-03 11 -155 val_155 2010-03-03 11 -113 val_113 2010-03-03 11 -17 val_17 2010-03-03 11 -475 val_475 2010-03-03 11 -195 val_195 2010-03-03 11 -167 val_167 2010-03-03 11 -325 val_325 2010-03-03 11 -367 val_367 2010-03-03 11 -342 val_342 2010-03-03 11 +119 val_119 2010-03-03 11 +414 val_414 2010-03-03 11 +200 val_200 2010-03-03 11 +491 val_491 2010-03-03 11 +237 val_237 2010-03-03 11 439 val_439 2010-03-03 11 -266 val_266 2010-03-03 11 -365 val_365 2010-03-03 11 -309 val_309 2010-03-03 11 -397 val_397 2010-03-03 11 -377 val_377 2010-03-03 11 -162 val_162 2010-03-03 11 -489 val_489 2010-03-03 11 +360 val_360 2010-03-03 11 +248 val_248 2010-03-03 11 +479 val_479 2010-03-03 11 +305 val_305 2010-03-03 11 417 val_417 2010-03-03 11 -247 val_247 2010-03-03 11 -396 val_396 2010-03-03 11 -399 val_399 2010-03-03 11 -174 val_174 2010-03-03 11 -208 val_208 2010-03-03 11 -466 val_466 2010-03-03 11 199 val_199 2010-03-03 11 -207 val_207 2010-03-03 11 -494 val_494 2010-03-03 11 -413 val_413 2010-03-03 11 -174 val_174 2010-03-03 11 -482 val_482 2010-03-03 11 -237 val_237 2010-03-03 11 -394 val_394 2010-03-03 11 -459 val_459 2010-03-03 11 -446 val_446 2010-03-03 11 -338 val_338 2010-03-03 11 -193 val_193 2010-03-03 11 -153 val_153 2010-03-03 11 -287 val_287 2010-03-03 11 -219 val_219 2010-03-03 11 -292 val_292 2010-03-03 11 -252 val_252 2010-03-03 11 -430 val_430 2010-03-03 11 -417 val_417 2010-03-03 11 -166 val_166 2010-03-03 11 -403 val_403 2010-03-03 11 -82 val_82 2010-03-03 11 -15 val_15 2010-03-03 11 -209 val_209 2010-03-03 11 -277 val_277 2010-03-03 11 -281 val_281 2010-03-03 11 -327 val_327 2010-03-03 11 -37 val_37 2010-03-03 11 -495 val_495 2010-03-03 11 -145 val_145 2010-03-03 11 -469 val_469 2010-03-03 11 -152 val_152 2010-03-03 11 -374 val_374 2010-03-03 11 +444 val_444 2010-03-03 11 +120 val_120 2010-03-03 11 429 val_429 2010-03-03 11 +169 val_169 2010-03-03 11 +443 val_443 2010-03-03 11 +323 val_323 2010-03-03 11 +325 val_325 2010-03-03 11 +277 val_277 2010-03-03 11 +230 val_230 2010-03-03 11 +478 val_478 2010-03-03 11 +178 val_178 2010-03-03 11 +468 val_468 2010-03-03 11 +310 val_310 2010-03-03 11 +317 val_317 2010-03-03 11 +333 val_333 2010-03-03 11 +493 val_493 2010-03-03 11 +460 val_460 2010-03-03 11 +207 val_207 2010-03-03 11 +249 val_249 2010-03-03 11 +265 val_265 2010-03-03 11 +480 val_480 2010-03-03 11 +83 val_83 2010-03-03 11 +136 val_136 2010-03-03 11 +353 val_353 2010-03-03 11 +172 val_172 2010-03-03 11 +214 val_214 2010-03-03 11 +462 val_462 2010-03-03 11 +233 val_233 2010-03-03 11 406 val_406 2010-03-03 11 +133 val_133 2010-03-03 11 +175 val_175 2010-03-03 11 +189 val_189 2010-03-03 11 +454 val_454 2010-03-03 11 +375 val_375 2010-03-03 11 +401 val_401 2010-03-03 11 +421 val_421 2010-03-03 11 +407 val_407 2010-03-03 11 +384 val_384 2010-03-03 11 +256 val_256 2010-03-03 11 +26 val_26 2010-03-03 11 +134 val_134 2010-03-03 11 +67 val_67 2010-03-03 11 +384 val_384 2010-03-03 11 +379 val_379 2010-03-03 11 +18 val_18 2010-03-03 11 +462 val_462 2010-03-03 11 +492 val_492 2010-03-03 11 +100 val_100 2010-03-03 11 +298 val_298 2010-03-03 11 +9 val_9 2010-03-03 11 +341 val_341 2010-03-03 11 +498 val_498 2010-03-03 11 146 val_146 2010-03-03 11 -213 val_213 2010-03-03 11 -128 val_128 2010-03-03 11 -66 val_66 2010-03-03 11 -369 val_369 2010-03-03 11 -224 val_224 2010-03-03 11 +458 val_458 2010-03-03 11 +362 val_362 2010-03-03 11 +186 val_186 2010-03-03 11 +285 val_285 2010-03-03 11 +348 val_348 2010-03-03 11 +167 val_167 2010-03-03 11 +18 val_18 2010-03-03 11 273 val_273 2010-03-03 11 -150 val_150 2010-03-03 11 -401 val_401 2010-03-03 11 -193 val_193 2010-03-03 11 -265 val_265 2010-03-03 11 -484 val_484 2010-03-03 11 -98 val_98 2010-03-03 11 -278 val_278 2010-03-03 11 -255 val_255 2010-03-03 11 -409 val_409 2010-03-03 11 -165 val_165 2010-03-03 11 -27 val_27 2010-03-03 11 -311 val_311 2010-03-03 11 -86 val_86 2010-03-03 11 -238 val_238 2010-03-03 11 +183 val_183 2010-03-03 11 +281 val_281 2010-03-03 11 +344 val_344 2010-03-03 11 +97 val_97 2010-03-03 11 +469 val_469 2010-03-03 11 +315 val_315 2010-03-03 11 +84 val_84 2010-03-03 11 +28 val_28 2010-03-03 11 +37 val_37 2010-03-03 11 +448 val_448 2010-03-03 11 +152 val_152 2010-03-03 11 +348 val_348 2010-03-03 11 +307 val_307 2010-03-03 11 +194 val_194 2010-03-03 11 +414 val_414 2010-03-03 11 +477 val_477 2010-03-03 11 +222 val_222 2010-03-03 11 +126 val_126 2010-03-03 11 +90 val_90 2010-03-03 11 +169 val_169 2010-03-03 11 +403 val_403 2010-03-03 11 +400 val_400 2010-03-03 11 +200 val_200 2010-03-03 11 +97 val_97 2010-03-03 11 PREHOOK: query: select * from nzhang_part6 where ds = '2010-03-03' and hr = '12' PREHOOK: type: QUERY PREHOOK: Input: default@nzhang_part6 @@ -1074,1003 +1074,1003 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@nzhang_part6 POSTHOOK: Input: default@nzhang_part6@ds=2010-03-03/hr=12 #### A masked pattern was here #### -97 val_97 2010-03-03 12 -200 val_200 2010-03-03 12 -400 val_400 2010-03-03 12 -403 val_403 2010-03-03 12 -169 val_169 2010-03-03 12 -90 val_90 2010-03-03 12 -126 val_126 2010-03-03 12 -222 val_222 2010-03-03 12 -477 val_477 2010-03-03 12 -414 val_414 2010-03-03 12 -194 val_194 2010-03-03 12 -307 val_307 2010-03-03 12 -348 val_348 2010-03-03 12 -152 val_152 2010-03-03 12 -448 val_448 2010-03-03 12 -37 val_37 2010-03-03 12 -28 val_28 2010-03-03 12 -84 val_84 2010-03-03 12 -315 val_315 2010-03-03 12 -469 val_469 2010-03-03 12 -97 val_97 2010-03-03 12 -344 val_344 2010-03-03 12 -281 val_281 2010-03-03 12 -183 val_183 2010-03-03 12 +238 val_238 2010-03-03 12 +86 val_86 2010-03-03 12 +311 val_311 2010-03-03 12 +27 val_27 2010-03-03 12 +165 val_165 2010-03-03 12 +409 val_409 2010-03-03 12 +255 val_255 2010-03-03 12 +278 val_278 2010-03-03 12 +98 val_98 2010-03-03 12 +484 val_484 2010-03-03 12 +265 val_265 2010-03-03 12 +193 val_193 2010-03-03 12 +401 val_401 2010-03-03 12 +150 val_150 2010-03-03 12 273 val_273 2010-03-03 12 -18 val_18 2010-03-03 12 -167 val_167 2010-03-03 12 -348 val_348 2010-03-03 12 -285 val_285 2010-03-03 12 -186 val_186 2010-03-03 12 -362 val_362 2010-03-03 12 -458 val_458 2010-03-03 12 +224 val_224 2010-03-03 12 +369 val_369 2010-03-03 12 +66 val_66 2010-03-03 12 +128 val_128 2010-03-03 12 +213 val_213 2010-03-03 12 146 val_146 2010-03-03 12 -498 val_498 2010-03-03 12 -341 val_341 2010-03-03 12 -9 val_9 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -492 val_492 2010-03-03 12 -462 val_462 2010-03-03 12 -18 val_18 2010-03-03 12 -379 val_379 2010-03-03 12 -384 val_384 2010-03-03 12 -67 val_67 2010-03-03 12 -134 val_134 2010-03-03 12 -26 val_26 2010-03-03 12 -256 val_256 2010-03-03 12 -384 val_384 2010-03-03 12 -407 val_407 2010-03-03 12 -421 val_421 2010-03-03 12 -401 val_401 2010-03-03 12 -375 val_375 2010-03-03 12 -454 val_454 2010-03-03 12 -189 val_189 2010-03-03 12 -175 val_175 2010-03-03 12 -133 val_133 2010-03-03 12 406 val_406 2010-03-03 12 -233 val_233 2010-03-03 12 -462 val_462 2010-03-03 12 -214 val_214 2010-03-03 12 -172 val_172 2010-03-03 12 -353 val_353 2010-03-03 12 -136 val_136 2010-03-03 12 -83 val_83 2010-03-03 12 -480 val_480 2010-03-03 12 -265 val_265 2010-03-03 12 -249 val_249 2010-03-03 12 -207 val_207 2010-03-03 12 -460 val_460 2010-03-03 12 -493 val_493 2010-03-03 12 -333 val_333 2010-03-03 12 -317 val_317 2010-03-03 12 -310 val_310 2010-03-03 12 -468 val_468 2010-03-03 12 -178 val_178 2010-03-03 12 -478 val_478 2010-03-03 12 -230 val_230 2010-03-03 12 -277 val_277 2010-03-03 12 -325 val_325 2010-03-03 12 -323 val_323 2010-03-03 12 -443 val_443 2010-03-03 12 -169 val_169 2010-03-03 12 429 val_429 2010-03-03 12 -120 val_120 2010-03-03 12 -444 val_444 2010-03-03 12 -199 val_199 2010-03-03 12 +374 val_374 2010-03-03 12 +152 val_152 2010-03-03 12 +469 val_469 2010-03-03 12 +145 val_145 2010-03-03 12 +495 val_495 2010-03-03 12 +37 val_37 2010-03-03 12 +327 val_327 2010-03-03 12 +281 val_281 2010-03-03 12 +277 val_277 2010-03-03 12 +209 val_209 2010-03-03 12 +15 val_15 2010-03-03 12 +82 val_82 2010-03-03 12 +403 val_403 2010-03-03 12 +166 val_166 2010-03-03 12 417 val_417 2010-03-03 12 -305 val_305 2010-03-03 12 -479 val_479 2010-03-03 12 -248 val_248 2010-03-03 12 -360 val_360 2010-03-03 12 -439 val_439 2010-03-03 12 +430 val_430 2010-03-03 12 +252 val_252 2010-03-03 12 +292 val_292 2010-03-03 12 +219 val_219 2010-03-03 12 +287 val_287 2010-03-03 12 +153 val_153 2010-03-03 12 +193 val_193 2010-03-03 12 +338 val_338 2010-03-03 12 +446 val_446 2010-03-03 12 +459 val_459 2010-03-03 12 +394 val_394 2010-03-03 12 237 val_237 2010-03-03 12 -491 val_491 2010-03-03 12 -200 val_200 2010-03-03 12 -414 val_414 2010-03-03 12 -119 val_119 2010-03-03 12 -438 val_438 2010-03-03 12 -163 val_163 2010-03-03 12 -70 val_70 2010-03-03 12 -104 val_104 2010-03-03 12 -255 val_255 2010-03-03 12 -351 val_351 2010-03-03 12 -24 val_24 2010-03-03 12 -291 val_291 2010-03-03 12 -480 val_480 2010-03-03 12 -397 val_397 2010-03-03 12 -70 val_70 2010-03-03 12 -5 val_5 2010-03-03 12 -382 val_382 2010-03-03 12 -187 val_187 2010-03-03 12 -424 val_424 2010-03-03 12 -164 val_164 2010-03-03 12 -431 val_431 2010-03-03 12 -125 val_125 2010-03-03 12 -298 val_298 2010-03-03 12 -478 val_478 2010-03-03 12 -454 val_454 2010-03-03 12 -431 val_431 2010-03-03 12 -164 val_164 2010-03-03 12 -217 val_217 2010-03-03 12 -201 val_201 2010-03-03 12 +482 val_482 2010-03-03 12 +174 val_174 2010-03-03 12 +413 val_413 2010-03-03 12 +494 val_494 2010-03-03 12 +207 val_207 2010-03-03 12 +199 val_199 2010-03-03 12 +466 val_466 2010-03-03 12 +208 val_208 2010-03-03 12 +174 val_174 2010-03-03 12 +399 val_399 2010-03-03 12 396 val_396 2010-03-03 12 -12 val_12 2010-03-03 12 -424 val_424 2010-03-03 12 -348 val_348 2010-03-03 12 -262 val_262 2010-03-03 12 +247 val_247 2010-03-03 12 +417 val_417 2010-03-03 12 +489 val_489 2010-03-03 12 +162 val_162 2010-03-03 12 +377 val_377 2010-03-03 12 +397 val_397 2010-03-03 12 +309 val_309 2010-03-03 12 +365 val_365 2010-03-03 12 +266 val_266 2010-03-03 12 +439 val_439 2010-03-03 12 +342 val_342 2010-03-03 12 +367 val_367 2010-03-03 12 +325 val_325 2010-03-03 12 +167 val_167 2010-03-03 12 +195 val_195 2010-03-03 12 +475 val_475 2010-03-03 12 +17 val_17 2010-03-03 12 +113 val_113 2010-03-03 12 +155 val_155 2010-03-03 12 203 val_203 2010-03-03 12 -90 val_90 2010-03-03 12 -258 val_258 2010-03-03 12 -114 val_114 2010-03-03 12 -401 val_401 2010-03-03 12 -406 val_406 2010-03-03 12 -190 val_190 2010-03-03 12 -409 val_409 2010-03-03 12 -406 val_406 2010-03-03 12 -257 val_257 2010-03-03 12 -105 val_105 2010-03-03 12 -53 val_53 2010-03-03 12 -483 val_483 2010-03-03 12 -403 val_403 2010-03-03 12 -175 val_175 2010-03-03 12 -366 val_366 2010-03-03 12 -466 val_466 2010-03-03 12 -104 val_104 2010-03-03 12 -335 val_335 2010-03-03 12 -321 val_321 2010-03-03 12 -193 val_193 2010-03-03 12 -44 val_44 2010-03-03 12 -80 val_80 2010-03-03 12 -235 val_235 2010-03-03 12 -331 val_331 2010-03-03 12 -283 val_283 2010-03-03 12 -35 val_35 2010-03-03 12 -2 val_2 2010-03-03 12 -280 val_280 2010-03-03 12 -463 val_463 2010-03-03 12 -469 val_469 2010-03-03 12 -229 val_229 2010-03-03 12 -316 val_316 2010-03-03 12 -202 val_202 2010-03-03 12 -432 val_432 2010-03-03 12 -467 val_467 2010-03-03 12 +339 val_339 2010-03-03 12 +0 val_0 2010-03-03 12 +455 val_455 2010-03-03 12 128 val_128 2010-03-03 12 +311 val_311 2010-03-03 12 +316 val_316 2010-03-03 12 +57 val_57 2010-03-03 12 +302 val_302 2010-03-03 12 +205 val_205 2010-03-03 12 +149 val_149 2010-03-03 12 438 val_438 2010-03-03 12 -244 val_244 2010-03-03 12 -5 val_5 2010-03-03 12 -191 val_191 2010-03-03 12 -288 val_288 2010-03-03 12 -401 val_401 2010-03-03 12 -480 val_480 2010-03-03 12 -487 val_487 2010-03-03 12 -70 val_70 2010-03-03 12 -263 val_263 2010-03-03 12 -256 val_256 2010-03-03 12 -223 val_223 2010-03-03 12 -116 val_116 2010-03-03 12 -485 val_485 2010-03-03 12 -239 val_239 2010-03-03 12 -219 val_219 2010-03-03 12 -274 val_274 2010-03-03 12 -167 val_167 2010-03-03 12 -344 val_344 2010-03-03 12 -367 val_367 2010-03-03 12 -216 val_216 2010-03-03 12 -113 val_113 2010-03-03 12 -296 val_296 2010-03-03 12 -103 val_103 2010-03-03 12 -368 val_368 2010-03-03 12 -33 val_33 2010-03-03 12 -230 val_230 2010-03-03 12 -69 val_69 2010-03-03 12 -342 val_342 2010-03-03 12 -74 val_74 2010-03-03 12 -76 val_76 2010-03-03 12 -468 val_468 2010-03-03 12 -64 val_64 2010-03-03 12 -209 val_209 2010-03-03 12 -30 val_30 2010-03-03 12 -453 val_453 2010-03-03 12 +345 val_345 2010-03-03 12 +129 val_129 2010-03-03 12 +170 val_170 2010-03-03 12 +20 val_20 2010-03-03 12 +489 val_489 2010-03-03 12 +157 val_157 2010-03-03 12 +378 val_378 2010-03-03 12 +221 val_221 2010-03-03 12 +92 val_92 2010-03-03 12 +111 val_111 2010-03-03 12 +47 val_47 2010-03-03 12 +72 val_72 2010-03-03 12 +4 val_4 2010-03-03 12 +280 val_280 2010-03-03 12 +35 val_35 2010-03-03 12 +427 val_427 2010-03-03 12 +277 val_277 2010-03-03 12 +208 val_208 2010-03-03 12 +356 val_356 2010-03-03 12 +399 val_399 2010-03-03 12 +169 val_169 2010-03-03 12 +382 val_382 2010-03-03 12 +498 val_498 2010-03-03 12 +125 val_125 2010-03-03 12 +386 val_386 2010-03-03 12 +437 val_437 2010-03-03 12 +469 val_469 2010-03-03 12 +192 val_192 2010-03-03 12 +286 val_286 2010-03-03 12 +187 val_187 2010-03-03 12 +176 val_176 2010-03-03 12 +54 val_54 2010-03-03 12 +459 val_459 2010-03-03 12 +51 val_51 2010-03-03 12 138 val_138 2010-03-03 12 -228 val_228 2010-03-03 12 -218 val_218 2010-03-03 12 -449 val_449 2010-03-03 12 -149 val_149 2010-03-03 12 -492 val_492 2010-03-03 12 -223 val_223 2010-03-03 12 -41 val_41 2010-03-03 12 -76 val_76 2010-03-03 12 -78 val_78 2010-03-03 12 -458 val_458 2010-03-03 12 -489 val_489 2010-03-03 12 -119 val_119 2010-03-03 12 +103 val_103 2010-03-03 12 +239 val_239 2010-03-03 12 +213 val_213 2010-03-03 12 +216 val_216 2010-03-03 12 430 val_430 2010-03-03 12 -321 val_321 2010-03-03 12 -42 val_42 2010-03-03 12 -195 val_195 2010-03-03 12 -160 val_160 2010-03-03 12 -498 val_498 2010-03-03 12 -322 val_322 2010-03-03 12 -472 val_472 2010-03-03 12 -143 val_143 2010-03-03 12 -233 val_233 2010-03-03 12 -229 val_229 2010-03-03 12 -34 val_34 2010-03-03 12 -168 val_168 2010-03-03 12 -11 val_11 2010-03-03 12 -95 val_95 2010-03-03 12 -336 val_336 2010-03-03 12 -35 val_35 2010-03-03 12 -58 val_58 2010-03-03 12 -395 val_395 2010-03-03 12 -317 val_317 2010-03-03 12 -396 val_396 2010-03-03 12 -402 val_402 2010-03-03 12 -497 val_497 2010-03-03 12 -5 val_5 2010-03-03 12 -226 val_226 2010-03-03 12 -177 val_177 2010-03-03 12 -452 val_452 2010-03-03 12 -242 val_242 2010-03-03 12 -401 val_401 2010-03-03 12 -331 val_331 2010-03-03 12 +278 val_278 2010-03-03 12 +176 val_176 2010-03-03 12 +289 val_289 2010-03-03 12 +221 val_221 2010-03-03 12 +65 val_65 2010-03-03 12 +318 val_318 2010-03-03 12 +332 val_332 2010-03-03 12 +311 val_311 2010-03-03 12 +275 val_275 2010-03-03 12 +137 val_137 2010-03-03 12 +241 val_241 2010-03-03 12 +83 val_83 2010-03-03 12 +333 val_333 2010-03-03 12 +180 val_180 2010-03-03 12 +284 val_284 2010-03-03 12 +12 val_12 2010-03-03 12 +230 val_230 2010-03-03 12 +181 val_181 2010-03-03 12 +67 val_67 2010-03-03 12 +260 val_260 2010-03-03 12 +404 val_404 2010-03-03 12 +384 val_384 2010-03-03 12 +489 val_489 2010-03-03 12 +353 val_353 2010-03-03 12 +373 val_373 2010-03-03 12 272 val_272 2010-03-03 12 -392 val_392 2010-03-03 12 -369 val_369 2010-03-03 12 -242 val_242 2010-03-03 12 -327 val_327 2010-03-03 12 -389 val_389 2010-03-03 12 -309 val_309 2010-03-03 12 -224 val_224 2010-03-03 12 -306 val_306 2010-03-03 12 -273 val_273 2010-03-03 12 -277 val_277 2010-03-03 12 -10 val_10 2010-03-03 12 -435 val_435 2010-03-03 12 -19 val_19 2010-03-03 12 -307 val_307 2010-03-03 12 -90 val_90 2010-03-03 12 -72 val_72 2010-03-03 12 -118 val_118 2010-03-03 12 -15 val_15 2010-03-03 12 -419 val_419 2010-03-03 12 -238 val_238 2010-03-03 12 138 val_138 2010-03-03 12 -282 val_282 2010-03-03 12 -395 val_395 2010-03-03 12 -134 val_134 2010-03-03 12 -118 val_118 2010-03-03 12 +217 val_217 2010-03-03 12 +84 val_84 2010-03-03 12 +348 val_348 2010-03-03 12 +466 val_466 2010-03-03 12 +58 val_58 2010-03-03 12 +8 val_8 2010-03-03 12 +411 val_411 2010-03-03 12 +230 val_230 2010-03-03 12 +208 val_208 2010-03-03 12 +348 val_348 2010-03-03 12 +24 val_24 2010-03-03 12 +463 val_463 2010-03-03 12 +431 val_431 2010-03-03 12 179 val_179 2010-03-03 12 -364 val_364 2010-03-03 12 -87 val_87 2010-03-03 12 -490 val_490 2010-03-03 12 +172 val_172 2010-03-03 12 +42 val_42 2010-03-03 12 +129 val_129 2010-03-03 12 +158 val_158 2010-03-03 12 +119 val_119 2010-03-03 12 +496 val_496 2010-03-03 12 0 val_0 2010-03-03 12 -77 val_77 2010-03-03 12 -85 val_85 2010-03-03 12 -413 val_413 2010-03-03 12 -169 val_169 2010-03-03 12 -316 val_316 2010-03-03 12 -369 val_369 2010-03-03 12 -137 val_137 2010-03-03 12 -470 val_470 2010-03-03 12 -409 val_409 2010-03-03 12 -318 val_318 2010-03-03 12 -318 val_318 2010-03-03 12 -187 val_187 2010-03-03 12 +322 val_322 2010-03-03 12 197 val_197 2010-03-03 12 -282 val_282 2010-03-03 12 -98 val_98 2010-03-03 12 -457 val_457 2010-03-03 12 -481 val_481 2010-03-03 12 -288 val_288 2010-03-03 12 -196 val_196 2010-03-03 12 -95 val_95 2010-03-03 12 -308 val_308 2010-03-03 12 468 val_468 2010-03-03 12 -469 val_469 2010-03-03 12 -156 val_156 2010-03-03 12 -436 val_436 2010-03-03 12 -43 val_43 2010-03-03 12 -404 val_404 2010-03-03 12 -51 val_51 2010-03-03 12 -131 val_131 2010-03-03 12 -120 val_120 2010-03-03 12 -205 val_205 2010-03-03 12 -230 val_230 2010-03-03 12 -327 val_327 2010-03-03 12 -165 val_165 2010-03-03 12 -26 val_26 2010-03-03 12 -96 val_96 2010-03-03 12 -418 val_418 2010-03-03 12 -191 val_191 2010-03-03 12 -199 val_199 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -454 val_454 2010-03-03 12 393 val_393 2010-03-03 12 +454 val_454 2010-03-03 12 +100 val_100 2010-03-03 12 +298 val_298 2010-03-03 12 +199 val_199 2010-03-03 12 +191 val_191 2010-03-03 12 +418 val_418 2010-03-03 12 +96 val_96 2010-03-03 12 +26 val_26 2010-03-03 12 +165 val_165 2010-03-03 12 +327 val_327 2010-03-03 12 +230 val_230 2010-03-03 12 +205 val_205 2010-03-03 12 +120 val_120 2010-03-03 12 +131 val_131 2010-03-03 12 +51 val_51 2010-03-03 12 +404 val_404 2010-03-03 12 +43 val_43 2010-03-03 12 +436 val_436 2010-03-03 12 +156 val_156 2010-03-03 12 +469 val_469 2010-03-03 12 468 val_468 2010-03-03 12 +308 val_308 2010-03-03 12 +95 val_95 2010-03-03 12 +196 val_196 2010-03-03 12 +288 val_288 2010-03-03 12 +481 val_481 2010-03-03 12 +457 val_457 2010-03-03 12 +98 val_98 2010-03-03 12 +282 val_282 2010-03-03 12 197 val_197 2010-03-03 12 +187 val_187 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +409 val_409 2010-03-03 12 +470 val_470 2010-03-03 12 +137 val_137 2010-03-03 12 +369 val_369 2010-03-03 12 +316 val_316 2010-03-03 12 +169 val_169 2010-03-03 12 +413 val_413 2010-03-03 12 +85 val_85 2010-03-03 12 +77 val_77 2010-03-03 12 +0 val_0 2010-03-03 12 +490 val_490 2010-03-03 12 +87 val_87 2010-03-03 12 +364 val_364 2010-03-03 12 +179 val_179 2010-03-03 12 +118 val_118 2010-03-03 12 +134 val_134 2010-03-03 12 +395 val_395 2010-03-03 12 +282 val_282 2010-03-03 12 +138 val_138 2010-03-03 12 +238 val_238 2010-03-03 12 +419 val_419 2010-03-03 12 +15 val_15 2010-03-03 12 +118 val_118 2010-03-03 12 +72 val_72 2010-03-03 12 +90 val_90 2010-03-03 12 +307 val_307 2010-03-03 12 +19 val_19 2010-03-03 12 +435 val_435 2010-03-03 12 +10 val_10 2010-03-03 12 +277 val_277 2010-03-03 12 +273 val_273 2010-03-03 12 +306 val_306 2010-03-03 12 +224 val_224 2010-03-03 12 +309 val_309 2010-03-03 12 +389 val_389 2010-03-03 12 +327 val_327 2010-03-03 12 +242 val_242 2010-03-03 12 +369 val_369 2010-03-03 12 +392 val_392 2010-03-03 12 +272 val_272 2010-03-03 12 +331 val_331 2010-03-03 12 +401 val_401 2010-03-03 12 +242 val_242 2010-03-03 12 +452 val_452 2010-03-03 12 +177 val_177 2010-03-03 12 +226 val_226 2010-03-03 12 +5 val_5 2010-03-03 12 +497 val_497 2010-03-03 12 +402 val_402 2010-03-03 12 +396 val_396 2010-03-03 12 +317 val_317 2010-03-03 12 +395 val_395 2010-03-03 12 +58 val_58 2010-03-03 12 +35 val_35 2010-03-03 12 +336 val_336 2010-03-03 12 +95 val_95 2010-03-03 12 +11 val_11 2010-03-03 12 +168 val_168 2010-03-03 12 +34 val_34 2010-03-03 12 +229 val_229 2010-03-03 12 +233 val_233 2010-03-03 12 +143 val_143 2010-03-03 12 +472 val_472 2010-03-03 12 322 val_322 2010-03-03 12 -0 val_0 2010-03-03 12 -496 val_496 2010-03-03 12 -119 val_119 2010-03-03 12 -158 val_158 2010-03-03 12 -129 val_129 2010-03-03 12 +498 val_498 2010-03-03 12 +160 val_160 2010-03-03 12 +195 val_195 2010-03-03 12 42 val_42 2010-03-03 12 -172 val_172 2010-03-03 12 -179 val_179 2010-03-03 12 -431 val_431 2010-03-03 12 -463 val_463 2010-03-03 12 -24 val_24 2010-03-03 12 -348 val_348 2010-03-03 12 -208 val_208 2010-03-03 12 -230 val_230 2010-03-03 12 -411 val_411 2010-03-03 12 -8 val_8 2010-03-03 12 -58 val_58 2010-03-03 12 -466 val_466 2010-03-03 12 -348 val_348 2010-03-03 12 -84 val_84 2010-03-03 12 -217 val_217 2010-03-03 12 -138 val_138 2010-03-03 12 -272 val_272 2010-03-03 12 -373 val_373 2010-03-03 12 -353 val_353 2010-03-03 12 +321 val_321 2010-03-03 12 +430 val_430 2010-03-03 12 +119 val_119 2010-03-03 12 489 val_489 2010-03-03 12 -384 val_384 2010-03-03 12 -404 val_404 2010-03-03 12 -260 val_260 2010-03-03 12 -67 val_67 2010-03-03 12 -181 val_181 2010-03-03 12 +458 val_458 2010-03-03 12 +78 val_78 2010-03-03 12 +76 val_76 2010-03-03 12 +41 val_41 2010-03-03 12 +223 val_223 2010-03-03 12 +492 val_492 2010-03-03 12 +149 val_149 2010-03-03 12 +449 val_449 2010-03-03 12 +218 val_218 2010-03-03 12 +228 val_228 2010-03-03 12 +138 val_138 2010-03-03 12 +453 val_453 2010-03-03 12 +30 val_30 2010-03-03 12 +209 val_209 2010-03-03 12 +64 val_64 2010-03-03 12 +468 val_468 2010-03-03 12 +76 val_76 2010-03-03 12 +74 val_74 2010-03-03 12 +342 val_342 2010-03-03 12 +69 val_69 2010-03-03 12 230 val_230 2010-03-03 12 -12 val_12 2010-03-03 12 -284 val_284 2010-03-03 12 -180 val_180 2010-03-03 12 -333 val_333 2010-03-03 12 -83 val_83 2010-03-03 12 -241 val_241 2010-03-03 12 -137 val_137 2010-03-03 12 -275 val_275 2010-03-03 12 -311 val_311 2010-03-03 12 -332 val_332 2010-03-03 12 -318 val_318 2010-03-03 12 -65 val_65 2010-03-03 12 -221 val_221 2010-03-03 12 -289 val_289 2010-03-03 12 -176 val_176 2010-03-03 12 -278 val_278 2010-03-03 12 -430 val_430 2010-03-03 12 +33 val_33 2010-03-03 12 +368 val_368 2010-03-03 12 +103 val_103 2010-03-03 12 +296 val_296 2010-03-03 12 +113 val_113 2010-03-03 12 216 val_216 2010-03-03 12 -213 val_213 2010-03-03 12 +367 val_367 2010-03-03 12 +344 val_344 2010-03-03 12 +167 val_167 2010-03-03 12 +274 val_274 2010-03-03 12 +219 val_219 2010-03-03 12 239 val_239 2010-03-03 12 -103 val_103 2010-03-03 12 -138 val_138 2010-03-03 12 -51 val_51 2010-03-03 12 -459 val_459 2010-03-03 12 -54 val_54 2010-03-03 12 -176 val_176 2010-03-03 12 -187 val_187 2010-03-03 12 -286 val_286 2010-03-03 12 -192 val_192 2010-03-03 12 -469 val_469 2010-03-03 12 -437 val_437 2010-03-03 12 -386 val_386 2010-03-03 12 -125 val_125 2010-03-03 12 -498 val_498 2010-03-03 12 -382 val_382 2010-03-03 12 -169 val_169 2010-03-03 12 -399 val_399 2010-03-03 12 -356 val_356 2010-03-03 12 -208 val_208 2010-03-03 12 -277 val_277 2010-03-03 12 -427 val_427 2010-03-03 12 -35 val_35 2010-03-03 12 -280 val_280 2010-03-03 12 -4 val_4 2010-03-03 12 -72 val_72 2010-03-03 12 -47 val_47 2010-03-03 12 -111 val_111 2010-03-03 12 -92 val_92 2010-03-03 12 -221 val_221 2010-03-03 12 -378 val_378 2010-03-03 12 -157 val_157 2010-03-03 12 -489 val_489 2010-03-03 12 -20 val_20 2010-03-03 12 -170 val_170 2010-03-03 12 -129 val_129 2010-03-03 12 -345 val_345 2010-03-03 12 +485 val_485 2010-03-03 12 +116 val_116 2010-03-03 12 +223 val_223 2010-03-03 12 +256 val_256 2010-03-03 12 +263 val_263 2010-03-03 12 +70 val_70 2010-03-03 12 +487 val_487 2010-03-03 12 +480 val_480 2010-03-03 12 +401 val_401 2010-03-03 12 +288 val_288 2010-03-03 12 +191 val_191 2010-03-03 12 +5 val_5 2010-03-03 12 +244 val_244 2010-03-03 12 438 val_438 2010-03-03 12 -149 val_149 2010-03-03 12 -205 val_205 2010-03-03 12 -302 val_302 2010-03-03 12 -57 val_57 2010-03-03 12 -316 val_316 2010-03-03 12 -311 val_311 2010-03-03 12 128 val_128 2010-03-03 12 -455 val_455 2010-03-03 12 -0 val_0 2010-03-03 12 -339 val_339 2010-03-03 12 +467 val_467 2010-03-03 12 +432 val_432 2010-03-03 12 +202 val_202 2010-03-03 12 +316 val_316 2010-03-03 12 +229 val_229 2010-03-03 12 +469 val_469 2010-03-03 12 +463 val_463 2010-03-03 12 +280 val_280 2010-03-03 12 +2 val_2 2010-03-03 12 +35 val_35 2010-03-03 12 +283 val_283 2010-03-03 12 +331 val_331 2010-03-03 12 +235 val_235 2010-03-03 12 +80 val_80 2010-03-03 12 +44 val_44 2010-03-03 12 +193 val_193 2010-03-03 12 +321 val_321 2010-03-03 12 +335 val_335 2010-03-03 12 +104 val_104 2010-03-03 12 +466 val_466 2010-03-03 12 +366 val_366 2010-03-03 12 +175 val_175 2010-03-03 12 +403 val_403 2010-03-03 12 +483 val_483 2010-03-03 12 +53 val_53 2010-03-03 12 +105 val_105 2010-03-03 12 +257 val_257 2010-03-03 12 +406 val_406 2010-03-03 12 +409 val_409 2010-03-03 12 +190 val_190 2010-03-03 12 +406 val_406 2010-03-03 12 +401 val_401 2010-03-03 12 +114 val_114 2010-03-03 12 +258 val_258 2010-03-03 12 +90 val_90 2010-03-03 12 203 val_203 2010-03-03 12 -155 val_155 2010-03-03 12 -113 val_113 2010-03-03 12 -17 val_17 2010-03-03 12 -475 val_475 2010-03-03 12 -195 val_195 2010-03-03 12 -167 val_167 2010-03-03 12 -325 val_325 2010-03-03 12 -367 val_367 2010-03-03 12 -342 val_342 2010-03-03 12 -439 val_439 2010-03-03 12 -266 val_266 2010-03-03 12 -365 val_365 2010-03-03 12 -309 val_309 2010-03-03 12 -397 val_397 2010-03-03 12 -377 val_377 2010-03-03 12 -162 val_162 2010-03-03 12 -489 val_489 2010-03-03 12 -417 val_417 2010-03-03 12 -247 val_247 2010-03-03 12 +262 val_262 2010-03-03 12 +348 val_348 2010-03-03 12 +424 val_424 2010-03-03 12 +12 val_12 2010-03-03 12 396 val_396 2010-03-03 12 -399 val_399 2010-03-03 12 -174 val_174 2010-03-03 12 -208 val_208 2010-03-03 12 -466 val_466 2010-03-03 12 -199 val_199 2010-03-03 12 -207 val_207 2010-03-03 12 -494 val_494 2010-03-03 12 -413 val_413 2010-03-03 12 -174 val_174 2010-03-03 12 -482 val_482 2010-03-03 12 +201 val_201 2010-03-03 12 +217 val_217 2010-03-03 12 +164 val_164 2010-03-03 12 +431 val_431 2010-03-03 12 +454 val_454 2010-03-03 12 +478 val_478 2010-03-03 12 +298 val_298 2010-03-03 12 +125 val_125 2010-03-03 12 +431 val_431 2010-03-03 12 +164 val_164 2010-03-03 12 +424 val_424 2010-03-03 12 +187 val_187 2010-03-03 12 +382 val_382 2010-03-03 12 +5 val_5 2010-03-03 12 +70 val_70 2010-03-03 12 +397 val_397 2010-03-03 12 +480 val_480 2010-03-03 12 +291 val_291 2010-03-03 12 +24 val_24 2010-03-03 12 +351 val_351 2010-03-03 12 +255 val_255 2010-03-03 12 +104 val_104 2010-03-03 12 +70 val_70 2010-03-03 12 +163 val_163 2010-03-03 12 +438 val_438 2010-03-03 12 +119 val_119 2010-03-03 12 +414 val_414 2010-03-03 12 +200 val_200 2010-03-03 12 +491 val_491 2010-03-03 12 237 val_237 2010-03-03 12 -394 val_394 2010-03-03 12 -459 val_459 2010-03-03 12 -446 val_446 2010-03-03 12 -338 val_338 2010-03-03 12 -193 val_193 2010-03-03 12 -153 val_153 2010-03-03 12 -287 val_287 2010-03-03 12 -219 val_219 2010-03-03 12 -292 val_292 2010-03-03 12 -252 val_252 2010-03-03 12 -430 val_430 2010-03-03 12 +439 val_439 2010-03-03 12 +360 val_360 2010-03-03 12 +248 val_248 2010-03-03 12 +479 val_479 2010-03-03 12 +305 val_305 2010-03-03 12 417 val_417 2010-03-03 12 -166 val_166 2010-03-03 12 -403 val_403 2010-03-03 12 -82 val_82 2010-03-03 12 -15 val_15 2010-03-03 12 -209 val_209 2010-03-03 12 -277 val_277 2010-03-03 12 -281 val_281 2010-03-03 12 -327 val_327 2010-03-03 12 -37 val_37 2010-03-03 12 -495 val_495 2010-03-03 12 -145 val_145 2010-03-03 12 -469 val_469 2010-03-03 12 -152 val_152 2010-03-03 12 -374 val_374 2010-03-03 12 +199 val_199 2010-03-03 12 +444 val_444 2010-03-03 12 +120 val_120 2010-03-03 12 429 val_429 2010-03-03 12 +169 val_169 2010-03-03 12 +443 val_443 2010-03-03 12 +323 val_323 2010-03-03 12 +325 val_325 2010-03-03 12 +277 val_277 2010-03-03 12 +230 val_230 2010-03-03 12 +478 val_478 2010-03-03 12 +178 val_178 2010-03-03 12 +468 val_468 2010-03-03 12 +310 val_310 2010-03-03 12 +317 val_317 2010-03-03 12 +333 val_333 2010-03-03 12 +493 val_493 2010-03-03 12 +460 val_460 2010-03-03 12 +207 val_207 2010-03-03 12 +249 val_249 2010-03-03 12 +265 val_265 2010-03-03 12 +480 val_480 2010-03-03 12 +83 val_83 2010-03-03 12 +136 val_136 2010-03-03 12 +353 val_353 2010-03-03 12 +172 val_172 2010-03-03 12 +214 val_214 2010-03-03 12 +462 val_462 2010-03-03 12 +233 val_233 2010-03-03 12 406 val_406 2010-03-03 12 -146 val_146 2010-03-03 12 -213 val_213 2010-03-03 12 -128 val_128 2010-03-03 12 -66 val_66 2010-03-03 12 -369 val_369 2010-03-03 12 -224 val_224 2010-03-03 12 -273 val_273 2010-03-03 12 -150 val_150 2010-03-03 12 +133 val_133 2010-03-03 12 +175 val_175 2010-03-03 12 +189 val_189 2010-03-03 12 +454 val_454 2010-03-03 12 +375 val_375 2010-03-03 12 401 val_401 2010-03-03 12 -193 val_193 2010-03-03 12 -265 val_265 2010-03-03 12 -484 val_484 2010-03-03 12 -98 val_98 2010-03-03 12 -278 val_278 2010-03-03 12 -255 val_255 2010-03-03 12 -409 val_409 2010-03-03 12 -165 val_165 2010-03-03 12 -27 val_27 2010-03-03 12 -311 val_311 2010-03-03 12 -86 val_86 2010-03-03 12 -238 val_238 2010-03-03 12 -200 val_200 2010-03-03 12 -400 val_400 2010-03-03 12 -403 val_403 2010-03-03 12 -169 val_169 2010-03-03 12 -90 val_90 2010-03-03 12 -126 val_126 2010-03-03 12 -222 val_222 2010-03-03 12 -477 val_477 2010-03-03 12 -414 val_414 2010-03-03 12 -194 val_194 2010-03-03 12 -307 val_307 2010-03-03 12 +421 val_421 2010-03-03 12 +407 val_407 2010-03-03 12 +384 val_384 2010-03-03 12 +256 val_256 2010-03-03 12 +26 val_26 2010-03-03 12 +134 val_134 2010-03-03 12 +67 val_67 2010-03-03 12 +384 val_384 2010-03-03 12 +379 val_379 2010-03-03 12 +18 val_18 2010-03-03 12 +462 val_462 2010-03-03 12 +492 val_492 2010-03-03 12 +100 val_100 2010-03-03 12 +298 val_298 2010-03-03 12 +9 val_9 2010-03-03 12 +341 val_341 2010-03-03 12 +498 val_498 2010-03-03 12 +146 val_146 2010-03-03 12 +458 val_458 2010-03-03 12 +362 val_362 2010-03-03 12 +186 val_186 2010-03-03 12 +285 val_285 2010-03-03 12 348 val_348 2010-03-03 12 -152 val_152 2010-03-03 12 -448 val_448 2010-03-03 12 -37 val_37 2010-03-03 12 -28 val_28 2010-03-03 12 -84 val_84 2010-03-03 12 -315 val_315 2010-03-03 12 +167 val_167 2010-03-03 12 +18 val_18 2010-03-03 12 +273 val_273 2010-03-03 12 +183 val_183 2010-03-03 12 +281 val_281 2010-03-03 12 +344 val_344 2010-03-03 12 +97 val_97 2010-03-03 12 469 val_469 2010-03-03 12 +315 val_315 2010-03-03 12 +84 val_84 2010-03-03 12 +28 val_28 2010-03-03 12 +37 val_37 2010-03-03 12 +448 val_448 2010-03-03 12 +152 val_152 2010-03-03 12 +348 val_348 2010-03-03 12 +307 val_307 2010-03-03 12 +194 val_194 2010-03-03 12 +414 val_414 2010-03-03 12 +477 val_477 2010-03-03 12 +222 val_222 2010-03-03 12 +126 val_126 2010-03-03 12 +90 val_90 2010-03-03 12 +169 val_169 2010-03-03 12 +403 val_403 2010-03-03 12 +400 val_400 2010-03-03 12 +200 val_200 2010-03-03 12 97 val_97 2010-03-03 12 -344 val_344 2010-03-03 12 -281 val_281 2010-03-03 12 -183 val_183 2010-03-03 12 +238 val_238 2010-03-03 12 +86 val_86 2010-03-03 12 +311 val_311 2010-03-03 12 +27 val_27 2010-03-03 12 +165 val_165 2010-03-03 12 +409 val_409 2010-03-03 12 +255 val_255 2010-03-03 12 +278 val_278 2010-03-03 12 +98 val_98 2010-03-03 12 +484 val_484 2010-03-03 12 +265 val_265 2010-03-03 12 +193 val_193 2010-03-03 12 +401 val_401 2010-03-03 12 +150 val_150 2010-03-03 12 273 val_273 2010-03-03 12 -18 val_18 2010-03-03 12 -167 val_167 2010-03-03 12 -348 val_348 2010-03-03 12 -285 val_285 2010-03-03 12 -186 val_186 2010-03-03 12 -362 val_362 2010-03-03 12 -458 val_458 2010-03-03 12 +224 val_224 2010-03-03 12 +369 val_369 2010-03-03 12 +66 val_66 2010-03-03 12 +128 val_128 2010-03-03 12 +213 val_213 2010-03-03 12 146 val_146 2010-03-03 12 -498 val_498 2010-03-03 12 -341 val_341 2010-03-03 12 -9 val_9 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -492 val_492 2010-03-03 12 -462 val_462 2010-03-03 12 -18 val_18 2010-03-03 12 -379 val_379 2010-03-03 12 -384 val_384 2010-03-03 12 -67 val_67 2010-03-03 12 -134 val_134 2010-03-03 12 -26 val_26 2010-03-03 12 -256 val_256 2010-03-03 12 -384 val_384 2010-03-03 12 -407 val_407 2010-03-03 12 -421 val_421 2010-03-03 12 -401 val_401 2010-03-03 12 -375 val_375 2010-03-03 12 -454 val_454 2010-03-03 12 -189 val_189 2010-03-03 12 -175 val_175 2010-03-03 12 -133 val_133 2010-03-03 12 406 val_406 2010-03-03 12 -233 val_233 2010-03-03 12 -462 val_462 2010-03-03 12 -214 val_214 2010-03-03 12 -172 val_172 2010-03-03 12 -353 val_353 2010-03-03 12 -136 val_136 2010-03-03 12 -83 val_83 2010-03-03 12 -480 val_480 2010-03-03 12 -265 val_265 2010-03-03 12 -249 val_249 2010-03-03 12 -207 val_207 2010-03-03 12 -460 val_460 2010-03-03 12 -493 val_493 2010-03-03 12 -333 val_333 2010-03-03 12 -317 val_317 2010-03-03 12 -310 val_310 2010-03-03 12 -468 val_468 2010-03-03 12 -178 val_178 2010-03-03 12 -478 val_478 2010-03-03 12 -230 val_230 2010-03-03 12 -277 val_277 2010-03-03 12 -325 val_325 2010-03-03 12 -323 val_323 2010-03-03 12 -443 val_443 2010-03-03 12 -169 val_169 2010-03-03 12 429 val_429 2010-03-03 12 -120 val_120 2010-03-03 12 -444 val_444 2010-03-03 12 +374 val_374 2010-03-03 12 +152 val_152 2010-03-03 12 +469 val_469 2010-03-03 12 +145 val_145 2010-03-03 12 +495 val_495 2010-03-03 12 +37 val_37 2010-03-03 12 +327 val_327 2010-03-03 12 +281 val_281 2010-03-03 12 +277 val_277 2010-03-03 12 +209 val_209 2010-03-03 12 +15 val_15 2010-03-03 12 +82 val_82 2010-03-03 12 +403 val_403 2010-03-03 12 +166 val_166 2010-03-03 12 +417 val_417 2010-03-03 12 +430 val_430 2010-03-03 12 +252 val_252 2010-03-03 12 +292 val_292 2010-03-03 12 +219 val_219 2010-03-03 12 +287 val_287 2010-03-03 12 +153 val_153 2010-03-03 12 +193 val_193 2010-03-03 12 +338 val_338 2010-03-03 12 +446 val_446 2010-03-03 12 +459 val_459 2010-03-03 12 +394 val_394 2010-03-03 12 +237 val_237 2010-03-03 12 +482 val_482 2010-03-03 12 +174 val_174 2010-03-03 12 +413 val_413 2010-03-03 12 +494 val_494 2010-03-03 12 +207 val_207 2010-03-03 12 199 val_199 2010-03-03 12 +466 val_466 2010-03-03 12 +208 val_208 2010-03-03 12 +174 val_174 2010-03-03 12 +399 val_399 2010-03-03 12 +396 val_396 2010-03-03 12 +247 val_247 2010-03-03 12 417 val_417 2010-03-03 12 -305 val_305 2010-03-03 12 -479 val_479 2010-03-03 12 -248 val_248 2010-03-03 12 -360 val_360 2010-03-03 12 +489 val_489 2010-03-03 12 +162 val_162 2010-03-03 12 +377 val_377 2010-03-03 12 +397 val_397 2010-03-03 12 +309 val_309 2010-03-03 12 +365 val_365 2010-03-03 12 +266 val_266 2010-03-03 12 439 val_439 2010-03-03 12 -237 val_237 2010-03-03 12 -491 val_491 2010-03-03 12 -200 val_200 2010-03-03 12 -414 val_414 2010-03-03 12 -119 val_119 2010-03-03 12 +342 val_342 2010-03-03 12 +367 val_367 2010-03-03 12 +325 val_325 2010-03-03 12 +167 val_167 2010-03-03 12 +195 val_195 2010-03-03 12 +475 val_475 2010-03-03 12 +17 val_17 2010-03-03 12 +113 val_113 2010-03-03 12 +155 val_155 2010-03-03 12 +203 val_203 2010-03-03 12 +339 val_339 2010-03-03 12 +0 val_0 2010-03-03 12 +455 val_455 2010-03-03 12 +128 val_128 2010-03-03 12 +311 val_311 2010-03-03 12 +316 val_316 2010-03-03 12 +57 val_57 2010-03-03 12 +302 val_302 2010-03-03 12 +205 val_205 2010-03-03 12 +149 val_149 2010-03-03 12 438 val_438 2010-03-03 12 -163 val_163 2010-03-03 12 -70 val_70 2010-03-03 12 -104 val_104 2010-03-03 12 -255 val_255 2010-03-03 12 -351 val_351 2010-03-03 12 -24 val_24 2010-03-03 12 -291 val_291 2010-03-03 12 -480 val_480 2010-03-03 12 -397 val_397 2010-03-03 12 -70 val_70 2010-03-03 12 -5 val_5 2010-03-03 12 +345 val_345 2010-03-03 12 +129 val_129 2010-03-03 12 +170 val_170 2010-03-03 12 +20 val_20 2010-03-03 12 +489 val_489 2010-03-03 12 +157 val_157 2010-03-03 12 +378 val_378 2010-03-03 12 +221 val_221 2010-03-03 12 +92 val_92 2010-03-03 12 +111 val_111 2010-03-03 12 +47 val_47 2010-03-03 12 +72 val_72 2010-03-03 12 +4 val_4 2010-03-03 12 +280 val_280 2010-03-03 12 +35 val_35 2010-03-03 12 +427 val_427 2010-03-03 12 +277 val_277 2010-03-03 12 +208 val_208 2010-03-03 12 +356 val_356 2010-03-03 12 +399 val_399 2010-03-03 12 +169 val_169 2010-03-03 12 382 val_382 2010-03-03 12 -187 val_187 2010-03-03 12 -424 val_424 2010-03-03 12 -164 val_164 2010-03-03 12 -431 val_431 2010-03-03 12 +498 val_498 2010-03-03 12 125 val_125 2010-03-03 12 -298 val_298 2010-03-03 12 -478 val_478 2010-03-03 12 -454 val_454 2010-03-03 12 -431 val_431 2010-03-03 12 -164 val_164 2010-03-03 12 -217 val_217 2010-03-03 12 -201 val_201 2010-03-03 12 -396 val_396 2010-03-03 12 -12 val_12 2010-03-03 12 -424 val_424 2010-03-03 12 -348 val_348 2010-03-03 12 -262 val_262 2010-03-03 12 -203 val_203 2010-03-03 12 -90 val_90 2010-03-03 12 -258 val_258 2010-03-03 12 -114 val_114 2010-03-03 12 -401 val_401 2010-03-03 12 -406 val_406 2010-03-03 12 -190 val_190 2010-03-03 12 -409 val_409 2010-03-03 12 -406 val_406 2010-03-03 12 -257 val_257 2010-03-03 12 -105 val_105 2010-03-03 12 -53 val_53 2010-03-03 12 -483 val_483 2010-03-03 12 -403 val_403 2010-03-03 12 -175 val_175 2010-03-03 12 -366 val_366 2010-03-03 12 -466 val_466 2010-03-03 12 -104 val_104 2010-03-03 12 -335 val_335 2010-03-03 12 -321 val_321 2010-03-03 12 -193 val_193 2010-03-03 12 -44 val_44 2010-03-03 12 -80 val_80 2010-03-03 12 -235 val_235 2010-03-03 12 -331 val_331 2010-03-03 12 -283 val_283 2010-03-03 12 -35 val_35 2010-03-03 12 -2 val_2 2010-03-03 12 -280 val_280 2010-03-03 12 -463 val_463 2010-03-03 12 +386 val_386 2010-03-03 12 +437 val_437 2010-03-03 12 469 val_469 2010-03-03 12 -229 val_229 2010-03-03 12 -316 val_316 2010-03-03 12 -202 val_202 2010-03-03 12 -432 val_432 2010-03-03 12 -467 val_467 2010-03-03 12 -128 val_128 2010-03-03 12 -438 val_438 2010-03-03 12 -244 val_244 2010-03-03 12 -5 val_5 2010-03-03 12 -191 val_191 2010-03-03 12 -288 val_288 2010-03-03 12 -401 val_401 2010-03-03 12 -480 val_480 2010-03-03 12 -487 val_487 2010-03-03 12 -70 val_70 2010-03-03 12 -263 val_263 2010-03-03 12 -256 val_256 2010-03-03 12 -223 val_223 2010-03-03 12 -116 val_116 2010-03-03 12 -485 val_485 2010-03-03 12 +192 val_192 2010-03-03 12 +286 val_286 2010-03-03 12 +187 val_187 2010-03-03 12 +176 val_176 2010-03-03 12 +54 val_54 2010-03-03 12 +459 val_459 2010-03-03 12 +51 val_51 2010-03-03 12 +138 val_138 2010-03-03 12 +103 val_103 2010-03-03 12 239 val_239 2010-03-03 12 -219 val_219 2010-03-03 12 -274 val_274 2010-03-03 12 -167 val_167 2010-03-03 12 -344 val_344 2010-03-03 12 -367 val_367 2010-03-03 12 +213 val_213 2010-03-03 12 216 val_216 2010-03-03 12 -113 val_113 2010-03-03 12 -296 val_296 2010-03-03 12 -103 val_103 2010-03-03 12 -368 val_368 2010-03-03 12 -33 val_33 2010-03-03 12 +430 val_430 2010-03-03 12 +278 val_278 2010-03-03 12 +176 val_176 2010-03-03 12 +289 val_289 2010-03-03 12 +221 val_221 2010-03-03 12 +65 val_65 2010-03-03 12 +318 val_318 2010-03-03 12 +332 val_332 2010-03-03 12 +311 val_311 2010-03-03 12 +275 val_275 2010-03-03 12 +137 val_137 2010-03-03 12 +241 val_241 2010-03-03 12 +83 val_83 2010-03-03 12 +333 val_333 2010-03-03 12 +180 val_180 2010-03-03 12 +284 val_284 2010-03-03 12 +12 val_12 2010-03-03 12 230 val_230 2010-03-03 12 -69 val_69 2010-03-03 12 -342 val_342 2010-03-03 12 -74 val_74 2010-03-03 12 -76 val_76 2010-03-03 12 -468 val_468 2010-03-03 12 -64 val_64 2010-03-03 12 -209 val_209 2010-03-03 12 -30 val_30 2010-03-03 12 -453 val_453 2010-03-03 12 -138 val_138 2010-03-03 12 -228 val_228 2010-03-03 12 -218 val_218 2010-03-03 12 -449 val_449 2010-03-03 12 -149 val_149 2010-03-03 12 -492 val_492 2010-03-03 12 -223 val_223 2010-03-03 12 -41 val_41 2010-03-03 12 -76 val_76 2010-03-03 12 -78 val_78 2010-03-03 12 -458 val_458 2010-03-03 12 +181 val_181 2010-03-03 12 +67 val_67 2010-03-03 12 +260 val_260 2010-03-03 12 +404 val_404 2010-03-03 12 +384 val_384 2010-03-03 12 489 val_489 2010-03-03 12 -119 val_119 2010-03-03 12 -430 val_430 2010-03-03 12 -321 val_321 2010-03-03 12 -42 val_42 2010-03-03 12 -195 val_195 2010-03-03 12 -160 val_160 2010-03-03 12 -498 val_498 2010-03-03 12 -322 val_322 2010-03-03 12 -472 val_472 2010-03-03 12 -143 val_143 2010-03-03 12 -233 val_233 2010-03-03 12 -229 val_229 2010-03-03 12 -34 val_34 2010-03-03 12 -168 val_168 2010-03-03 12 -11 val_11 2010-03-03 12 -95 val_95 2010-03-03 12 -336 val_336 2010-03-03 12 -35 val_35 2010-03-03 12 -58 val_58 2010-03-03 12 -395 val_395 2010-03-03 12 -317 val_317 2010-03-03 12 -396 val_396 2010-03-03 12 -402 val_402 2010-03-03 12 -497 val_497 2010-03-03 12 -5 val_5 2010-03-03 12 -226 val_226 2010-03-03 12 -177 val_177 2010-03-03 12 -452 val_452 2010-03-03 12 -242 val_242 2010-03-03 12 -401 val_401 2010-03-03 12 -331 val_331 2010-03-03 12 +353 val_353 2010-03-03 12 +373 val_373 2010-03-03 12 272 val_272 2010-03-03 12 -392 val_392 2010-03-03 12 -369 val_369 2010-03-03 12 -242 val_242 2010-03-03 12 -327 val_327 2010-03-03 12 -389 val_389 2010-03-03 12 -309 val_309 2010-03-03 12 -224 val_224 2010-03-03 12 -306 val_306 2010-03-03 12 -273 val_273 2010-03-03 12 -277 val_277 2010-03-03 12 -10 val_10 2010-03-03 12 -435 val_435 2010-03-03 12 -19 val_19 2010-03-03 12 -307 val_307 2010-03-03 12 -90 val_90 2010-03-03 12 -72 val_72 2010-03-03 12 -118 val_118 2010-03-03 12 -15 val_15 2010-03-03 12 -419 val_419 2010-03-03 12 -238 val_238 2010-03-03 12 138 val_138 2010-03-03 12 -282 val_282 2010-03-03 12 -395 val_395 2010-03-03 12 -134 val_134 2010-03-03 12 -118 val_118 2010-03-03 12 +217 val_217 2010-03-03 12 +84 val_84 2010-03-03 12 +348 val_348 2010-03-03 12 +466 val_466 2010-03-03 12 +58 val_58 2010-03-03 12 +8 val_8 2010-03-03 12 +411 val_411 2010-03-03 12 +230 val_230 2010-03-03 12 +208 val_208 2010-03-03 12 +348 val_348 2010-03-03 12 +24 val_24 2010-03-03 12 +463 val_463 2010-03-03 12 +431 val_431 2010-03-03 12 179 val_179 2010-03-03 12 -364 val_364 2010-03-03 12 -87 val_87 2010-03-03 12 -490 val_490 2010-03-03 12 +172 val_172 2010-03-03 12 +42 val_42 2010-03-03 12 +129 val_129 2010-03-03 12 +158 val_158 2010-03-03 12 +119 val_119 2010-03-03 12 +496 val_496 2010-03-03 12 0 val_0 2010-03-03 12 -77 val_77 2010-03-03 12 -85 val_85 2010-03-03 12 -413 val_413 2010-03-03 12 -169 val_169 2010-03-03 12 -316 val_316 2010-03-03 12 -369 val_369 2010-03-03 12 -137 val_137 2010-03-03 12 -470 val_470 2010-03-03 12 -409 val_409 2010-03-03 12 -318 val_318 2010-03-03 12 -318 val_318 2010-03-03 12 -187 val_187 2010-03-03 12 +322 val_322 2010-03-03 12 197 val_197 2010-03-03 12 -282 val_282 2010-03-03 12 -98 val_98 2010-03-03 12 -457 val_457 2010-03-03 12 -481 val_481 2010-03-03 12 -288 val_288 2010-03-03 12 -196 val_196 2010-03-03 12 -95 val_95 2010-03-03 12 -308 val_308 2010-03-03 12 468 val_468 2010-03-03 12 -469 val_469 2010-03-03 12 -156 val_156 2010-03-03 12 -436 val_436 2010-03-03 12 -43 val_43 2010-03-03 12 -404 val_404 2010-03-03 12 -51 val_51 2010-03-03 12 -131 val_131 2010-03-03 12 -120 val_120 2010-03-03 12 -205 val_205 2010-03-03 12 -230 val_230 2010-03-03 12 -327 val_327 2010-03-03 12 -165 val_165 2010-03-03 12 -26 val_26 2010-03-03 12 -96 val_96 2010-03-03 12 -418 val_418 2010-03-03 12 -191 val_191 2010-03-03 12 -199 val_199 2010-03-03 12 -298 val_298 2010-03-03 12 -100 val_100 2010-03-03 12 -454 val_454 2010-03-03 12 393 val_393 2010-03-03 12 +454 val_454 2010-03-03 12 +100 val_100 2010-03-03 12 +298 val_298 2010-03-03 12 +199 val_199 2010-03-03 12 +191 val_191 2010-03-03 12 +418 val_418 2010-03-03 12 +96 val_96 2010-03-03 12 +26 val_26 2010-03-03 12 +165 val_165 2010-03-03 12 +327 val_327 2010-03-03 12 +230 val_230 2010-03-03 12 +205 val_205 2010-03-03 12 +120 val_120 2010-03-03 12 +131 val_131 2010-03-03 12 +51 val_51 2010-03-03 12 +404 val_404 2010-03-03 12 +43 val_43 2010-03-03 12 +436 val_436 2010-03-03 12 +156 val_156 2010-03-03 12 +469 val_469 2010-03-03 12 468 val_468 2010-03-03 12 +308 val_308 2010-03-03 12 +95 val_95 2010-03-03 12 +196 val_196 2010-03-03 12 +288 val_288 2010-03-03 12 +481 val_481 2010-03-03 12 +457 val_457 2010-03-03 12 +98 val_98 2010-03-03 12 +282 val_282 2010-03-03 12 197 val_197 2010-03-03 12 -322 val_322 2010-03-03 12 +187 val_187 2010-03-03 12 +318 val_318 2010-03-03 12 +318 val_318 2010-03-03 12 +409 val_409 2010-03-03 12 +470 val_470 2010-03-03 12 +137 val_137 2010-03-03 12 +369 val_369 2010-03-03 12 +316 val_316 2010-03-03 12 +169 val_169 2010-03-03 12 +413 val_413 2010-03-03 12 +85 val_85 2010-03-03 12 +77 val_77 2010-03-03 12 0 val_0 2010-03-03 12 -496 val_496 2010-03-03 12 -119 val_119 2010-03-03 12 -158 val_158 2010-03-03 12 -129 val_129 2010-03-03 12 -42 val_42 2010-03-03 12 -172 val_172 2010-03-03 12 +490 val_490 2010-03-03 12 +87 val_87 2010-03-03 12 +364 val_364 2010-03-03 12 179 val_179 2010-03-03 12 -431 val_431 2010-03-03 12 -463 val_463 2010-03-03 12 -24 val_24 2010-03-03 12 -348 val_348 2010-03-03 12 -208 val_208 2010-03-03 12 -230 val_230 2010-03-03 12 -411 val_411 2010-03-03 12 -8 val_8 2010-03-03 12 -58 val_58 2010-03-03 12 -466 val_466 2010-03-03 12 -348 val_348 2010-03-03 12 -84 val_84 2010-03-03 12 -217 val_217 2010-03-03 12 +118 val_118 2010-03-03 12 +134 val_134 2010-03-03 12 +395 val_395 2010-03-03 12 +282 val_282 2010-03-03 12 138 val_138 2010-03-03 12 +238 val_238 2010-03-03 12 +419 val_419 2010-03-03 12 +15 val_15 2010-03-03 12 +118 val_118 2010-03-03 12 +72 val_72 2010-03-03 12 +90 val_90 2010-03-03 12 +307 val_307 2010-03-03 12 +19 val_19 2010-03-03 12 +435 val_435 2010-03-03 12 +10 val_10 2010-03-03 12 +277 val_277 2010-03-03 12 +273 val_273 2010-03-03 12 +306 val_306 2010-03-03 12 +224 val_224 2010-03-03 12 +309 val_309 2010-03-03 12 +389 val_389 2010-03-03 12 +327 val_327 2010-03-03 12 +242 val_242 2010-03-03 12 +369 val_369 2010-03-03 12 +392 val_392 2010-03-03 12 272 val_272 2010-03-03 12 -373 val_373 2010-03-03 12 -353 val_353 2010-03-03 12 +331 val_331 2010-03-03 12 +401 val_401 2010-03-03 12 +242 val_242 2010-03-03 12 +452 val_452 2010-03-03 12 +177 val_177 2010-03-03 12 +226 val_226 2010-03-03 12 +5 val_5 2010-03-03 12 +497 val_497 2010-03-03 12 +402 val_402 2010-03-03 12 +396 val_396 2010-03-03 12 +317 val_317 2010-03-03 12 +395 val_395 2010-03-03 12 +58 val_58 2010-03-03 12 +35 val_35 2010-03-03 12 +336 val_336 2010-03-03 12 +95 val_95 2010-03-03 12 +11 val_11 2010-03-03 12 +168 val_168 2010-03-03 12 +34 val_34 2010-03-03 12 +229 val_229 2010-03-03 12 +233 val_233 2010-03-03 12 +143 val_143 2010-03-03 12 +472 val_472 2010-03-03 12 +322 val_322 2010-03-03 12 +498 val_498 2010-03-03 12 +160 val_160 2010-03-03 12 +195 val_195 2010-03-03 12 +42 val_42 2010-03-03 12 +321 val_321 2010-03-03 12 +430 val_430 2010-03-03 12 +119 val_119 2010-03-03 12 489 val_489 2010-03-03 12 -384 val_384 2010-03-03 12 -404 val_404 2010-03-03 12 -260 val_260 2010-03-03 12 -67 val_67 2010-03-03 12 -181 val_181 2010-03-03 12 +458 val_458 2010-03-03 12 +78 val_78 2010-03-03 12 +76 val_76 2010-03-03 12 +41 val_41 2010-03-03 12 +223 val_223 2010-03-03 12 +492 val_492 2010-03-03 12 +149 val_149 2010-03-03 12 +449 val_449 2010-03-03 12 +218 val_218 2010-03-03 12 +228 val_228 2010-03-03 12 +138 val_138 2010-03-03 12 +453 val_453 2010-03-03 12 +30 val_30 2010-03-03 12 +209 val_209 2010-03-03 12 +64 val_64 2010-03-03 12 +468 val_468 2010-03-03 12 +76 val_76 2010-03-03 12 +74 val_74 2010-03-03 12 +342 val_342 2010-03-03 12 +69 val_69 2010-03-03 12 230 val_230 2010-03-03 12 -12 val_12 2010-03-03 12 -284 val_284 2010-03-03 12 -180 val_180 2010-03-03 12 -333 val_333 2010-03-03 12 -83 val_83 2010-03-03 12 -241 val_241 2010-03-03 12 -137 val_137 2010-03-03 12 -275 val_275 2010-03-03 12 -311 val_311 2010-03-03 12 -332 val_332 2010-03-03 12 -318 val_318 2010-03-03 12 -65 val_65 2010-03-03 12 -221 val_221 2010-03-03 12 -289 val_289 2010-03-03 12 -176 val_176 2010-03-03 12 -278 val_278 2010-03-03 12 -430 val_430 2010-03-03 12 +33 val_33 2010-03-03 12 +368 val_368 2010-03-03 12 +103 val_103 2010-03-03 12 +296 val_296 2010-03-03 12 +113 val_113 2010-03-03 12 216 val_216 2010-03-03 12 -213 val_213 2010-03-03 12 +367 val_367 2010-03-03 12 +344 val_344 2010-03-03 12 +167 val_167 2010-03-03 12 +274 val_274 2010-03-03 12 +219 val_219 2010-03-03 12 239 val_239 2010-03-03 12 -103 val_103 2010-03-03 12 -138 val_138 2010-03-03 12 -51 val_51 2010-03-03 12 -459 val_459 2010-03-03 12 -54 val_54 2010-03-03 12 -176 val_176 2010-03-03 12 -187 val_187 2010-03-03 12 -286 val_286 2010-03-03 12 -192 val_192 2010-03-03 12 +485 val_485 2010-03-03 12 +116 val_116 2010-03-03 12 +223 val_223 2010-03-03 12 +256 val_256 2010-03-03 12 +263 val_263 2010-03-03 12 +70 val_70 2010-03-03 12 +487 val_487 2010-03-03 12 +480 val_480 2010-03-03 12 +401 val_401 2010-03-03 12 +288 val_288 2010-03-03 12 +191 val_191 2010-03-03 12 +5 val_5 2010-03-03 12 +244 val_244 2010-03-03 12 +438 val_438 2010-03-03 12 +128 val_128 2010-03-03 12 +467 val_467 2010-03-03 12 +432 val_432 2010-03-03 12 +202 val_202 2010-03-03 12 +316 val_316 2010-03-03 12 +229 val_229 2010-03-03 12 469 val_469 2010-03-03 12 -437 val_437 2010-03-03 12 -386 val_386 2010-03-03 12 +463 val_463 2010-03-03 12 +280 val_280 2010-03-03 12 +2 val_2 2010-03-03 12 +35 val_35 2010-03-03 12 +283 val_283 2010-03-03 12 +331 val_331 2010-03-03 12 +235 val_235 2010-03-03 12 +80 val_80 2010-03-03 12 +44 val_44 2010-03-03 12 +193 val_193 2010-03-03 12 +321 val_321 2010-03-03 12 +335 val_335 2010-03-03 12 +104 val_104 2010-03-03 12 +466 val_466 2010-03-03 12 +366 val_366 2010-03-03 12 +175 val_175 2010-03-03 12 +403 val_403 2010-03-03 12 +483 val_483 2010-03-03 12 +53 val_53 2010-03-03 12 +105 val_105 2010-03-03 12 +257 val_257 2010-03-03 12 +406 val_406 2010-03-03 12 +409 val_409 2010-03-03 12 +190 val_190 2010-03-03 12 +406 val_406 2010-03-03 12 +401 val_401 2010-03-03 12 +114 val_114 2010-03-03 12 +258 val_258 2010-03-03 12 +90 val_90 2010-03-03 12 +203 val_203 2010-03-03 12 +262 val_262 2010-03-03 12 +348 val_348 2010-03-03 12 +424 val_424 2010-03-03 12 +12 val_12 2010-03-03 12 +396 val_396 2010-03-03 12 +201 val_201 2010-03-03 12 +217 val_217 2010-03-03 12 +164 val_164 2010-03-03 12 +431 val_431 2010-03-03 12 +454 val_454 2010-03-03 12 +478 val_478 2010-03-03 12 +298 val_298 2010-03-03 12 125 val_125 2010-03-03 12 -498 val_498 2010-03-03 12 +431 val_431 2010-03-03 12 +164 val_164 2010-03-03 12 +424 val_424 2010-03-03 12 +187 val_187 2010-03-03 12 382 val_382 2010-03-03 12 -169 val_169 2010-03-03 12 -399 val_399 2010-03-03 12 -356 val_356 2010-03-03 12 -208 val_208 2010-03-03 12 -277 val_277 2010-03-03 12 -427 val_427 2010-03-03 12 -35 val_35 2010-03-03 12 -280 val_280 2010-03-03 12 -4 val_4 2010-03-03 12 -72 val_72 2010-03-03 12 -47 val_47 2010-03-03 12 -111 val_111 2010-03-03 12 -92 val_92 2010-03-03 12 -221 val_221 2010-03-03 12 -378 val_378 2010-03-03 12 -157 val_157 2010-03-03 12 -489 val_489 2010-03-03 12 -20 val_20 2010-03-03 12 -170 val_170 2010-03-03 12 -129 val_129 2010-03-03 12 -345 val_345 2010-03-03 12 +5 val_5 2010-03-03 12 +70 val_70 2010-03-03 12 +397 val_397 2010-03-03 12 +480 val_480 2010-03-03 12 +291 val_291 2010-03-03 12 +24 val_24 2010-03-03 12 +351 val_351 2010-03-03 12 +255 val_255 2010-03-03 12 +104 val_104 2010-03-03 12 +70 val_70 2010-03-03 12 +163 val_163 2010-03-03 12 438 val_438 2010-03-03 12 -149 val_149 2010-03-03 12 -205 val_205 2010-03-03 12 -302 val_302 2010-03-03 12 -57 val_57 2010-03-03 12 -316 val_316 2010-03-03 12 -311 val_311 2010-03-03 12 -128 val_128 2010-03-03 12 -455 val_455 2010-03-03 12 -0 val_0 2010-03-03 12 -339 val_339 2010-03-03 12 -203 val_203 2010-03-03 12 -155 val_155 2010-03-03 12 -113 val_113 2010-03-03 12 -17 val_17 2010-03-03 12 -475 val_475 2010-03-03 12 -195 val_195 2010-03-03 12 -167 val_167 2010-03-03 12 -325 val_325 2010-03-03 12 -367 val_367 2010-03-03 12 -342 val_342 2010-03-03 12 +119 val_119 2010-03-03 12 +414 val_414 2010-03-03 12 +200 val_200 2010-03-03 12 +491 val_491 2010-03-03 12 +237 val_237 2010-03-03 12 439 val_439 2010-03-03 12 -266 val_266 2010-03-03 12 -365 val_365 2010-03-03 12 -309 val_309 2010-03-03 12 -397 val_397 2010-03-03 12 -377 val_377 2010-03-03 12 -162 val_162 2010-03-03 12 -489 val_489 2010-03-03 12 +360 val_360 2010-03-03 12 +248 val_248 2010-03-03 12 +479 val_479 2010-03-03 12 +305 val_305 2010-03-03 12 417 val_417 2010-03-03 12 -247 val_247 2010-03-03 12 -396 val_396 2010-03-03 12 -399 val_399 2010-03-03 12 -174 val_174 2010-03-03 12 -208 val_208 2010-03-03 12 -466 val_466 2010-03-03 12 199 val_199 2010-03-03 12 -207 val_207 2010-03-03 12 -494 val_494 2010-03-03 12 -413 val_413 2010-03-03 12 -174 val_174 2010-03-03 12 -482 val_482 2010-03-03 12 -237 val_237 2010-03-03 12 -394 val_394 2010-03-03 12 -459 val_459 2010-03-03 12 -446 val_446 2010-03-03 12 -338 val_338 2010-03-03 12 -193 val_193 2010-03-03 12 -153 val_153 2010-03-03 12 -287 val_287 2010-03-03 12 -219 val_219 2010-03-03 12 -292 val_292 2010-03-03 12 -252 val_252 2010-03-03 12 -430 val_430 2010-03-03 12 -417 val_417 2010-03-03 12 -166 val_166 2010-03-03 12 -403 val_403 2010-03-03 12 -82 val_82 2010-03-03 12 -15 val_15 2010-03-03 12 -209 val_209 2010-03-03 12 -277 val_277 2010-03-03 12 -281 val_281 2010-03-03 12 -327 val_327 2010-03-03 12 -37 val_37 2010-03-03 12 -495 val_495 2010-03-03 12 -145 val_145 2010-03-03 12 -469 val_469 2010-03-03 12 -152 val_152 2010-03-03 12 -374 val_374 2010-03-03 12 +444 val_444 2010-03-03 12 +120 val_120 2010-03-03 12 429 val_429 2010-03-03 12 +169 val_169 2010-03-03 12 +443 val_443 2010-03-03 12 +323 val_323 2010-03-03 12 +325 val_325 2010-03-03 12 +277 val_277 2010-03-03 12 +230 val_230 2010-03-03 12 +478 val_478 2010-03-03 12 +178 val_178 2010-03-03 12 +468 val_468 2010-03-03 12 +310 val_310 2010-03-03 12 +317 val_317 2010-03-03 12 +333 val_333 2010-03-03 12 +493 val_493 2010-03-03 12 +460 val_460 2010-03-03 12 +207 val_207 2010-03-03 12 +249 val_249 2010-03-03 12 +265 val_265 2010-03-03 12 +480 val_480 2010-03-03 12 +83 val_83 2010-03-03 12 +136 val_136 2010-03-03 12 +353 val_353 2010-03-03 12 +172 val_172 2010-03-03 12 +214 val_214 2010-03-03 12 +462 val_462 2010-03-03 12 +233 val_233 2010-03-03 12 406 val_406 2010-03-03 12 +133 val_133 2010-03-03 12 +175 val_175 2010-03-03 12 +189 val_189 2010-03-03 12 +454 val_454 2010-03-03 12 +375 val_375 2010-03-03 12 +401 val_401 2010-03-03 12 +421 val_421 2010-03-03 12 +407 val_407 2010-03-03 12 +384 val_384 2010-03-03 12 +256 val_256 2010-03-03 12 +26 val_26 2010-03-03 12 +134 val_134 2010-03-03 12 +67 val_67 2010-03-03 12 +384 val_384 2010-03-03 12 +379 val_379 2010-03-03 12 +18 val_18 2010-03-03 12 +462 val_462 2010-03-03 12 +492 val_492 2010-03-03 12 +100 val_100 2010-03-03 12 +298 val_298 2010-03-03 12 +9 val_9 2010-03-03 12 +341 val_341 2010-03-03 12 +498 val_498 2010-03-03 12 146 val_146 2010-03-03 12 -213 val_213 2010-03-03 12 -128 val_128 2010-03-03 12 -66 val_66 2010-03-03 12 -369 val_369 2010-03-03 12 -224 val_224 2010-03-03 12 +458 val_458 2010-03-03 12 +362 val_362 2010-03-03 12 +186 val_186 2010-03-03 12 +285 val_285 2010-03-03 12 +348 val_348 2010-03-03 12 +167 val_167 2010-03-03 12 +18 val_18 2010-03-03 12 273 val_273 2010-03-03 12 -150 val_150 2010-03-03 12 -401 val_401 2010-03-03 12 -193 val_193 2010-03-03 12 -265 val_265 2010-03-03 12 -484 val_484 2010-03-03 12 -98 val_98 2010-03-03 12 -278 val_278 2010-03-03 12 -255 val_255 2010-03-03 12 -409 val_409 2010-03-03 12 -165 val_165 2010-03-03 12 -27 val_27 2010-03-03 12 -311 val_311 2010-03-03 12 -86 val_86 2010-03-03 12 -238 val_238 2010-03-03 12 +183 val_183 2010-03-03 12 +281 val_281 2010-03-03 12 +344 val_344 2010-03-03 12 +97 val_97 2010-03-03 12 +469 val_469 2010-03-03 12 +315 val_315 2010-03-03 12 +84 val_84 2010-03-03 12 +28 val_28 2010-03-03 12 +37 val_37 2010-03-03 12 +448 val_448 2010-03-03 12 +152 val_152 2010-03-03 12 +348 val_348 2010-03-03 12 +307 val_307 2010-03-03 12 +194 val_194 2010-03-03 12 +414 val_414 2010-03-03 12 +477 val_477 2010-03-03 12 +222 val_222 2010-03-03 12 +126 val_126 2010-03-03 12 +90 val_90 2010-03-03 12 +169 val_169 2010-03-03 12 +403 val_403 2010-03-03 12 +400 val_400 2010-03-03 12 +200 val_200 2010-03-03 12 97 val_97 2010-03-03 12 diff --git ql/src/test/results/clientpositive/load_dyn_part8.q.out ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out similarity index 73% rename from ql/src/test/results/clientpositive/load_dyn_part8.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out index a577b5b9b6..28e12fd289 100644 --- ql/src/test/results/clientpositive/load_dyn_part8.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out @@ -57,391 +57,435 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-0 depends on stages: Stage-2 - Stage-3 depends on stages: Stage-0 - Stage-1 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-1, Stage-5 - Stage-5 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-3 + Stage-4 depends on stages: Stage-0 + Stage-1 depends on stages: Stage-3 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: (ds <= '2008-04-08') (type: boolean) - Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Tez #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: (ds <= '2008-04-08') (type: boolean) + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 #### A masked pattern was here #### - name default.nzhang_part8_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Filter Operator - isSamplingPred: false - predicate: (ds > '2008-04-08') (type: boolean) - Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 2 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-12-31/ - Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE + name default.nzhang_part8_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Filter Operator + isSamplingPred: false + predicate: (ds > '2008-04-08') (type: boolean) + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 2 #### A masked pattern was here #### - name default.nzhang_part8_n0 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-12-31/ + Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8_n0 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, hr - Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: '2008-12-31' (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 1 - compressed: false - GlobalTableId: 0 + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: + name default.nzhang_part8_n0 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8_n0 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - Path -> Partition: + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8_n0 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, hr + Statistics: Num rows: 666 Data size: 303696 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: '2008-12-31' (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: '2008-12-31' (type: string), _col1 (type: string) + null sort order: zz + numBuckets: -1 + sort order: ++ + Map-reduce partition columns: '2008-12-31' (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col2 (type: struct), _col3 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + Path -> Partition: #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-09 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-09 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart #### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 #### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /srcpart/ds=2008-04-08/hr=11 [srcpart] - /srcpart/ds=2008-04-08/hr=12 [srcpart] - /srcpart/ds=2008-04-09/hr=11 [srcpart] - /srcpart/ds=2008-04-09/hr=12 [srcpart] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + /srcpart/ds=2008-04-09/hr=11 [srcpart] + /srcpart/ds=2008-04-09/hr=12 [srcpart] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + Reducer 3 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: '2008-12-31' (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), '2008-12-31' (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types struct:struct:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-3 + Dependency Collection Stage: Stage-0 Move Operator @@ -472,7 +516,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 - Stage: Stage-3 + Stage: Stage-4 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -506,7 +550,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 - Stage: Stage-4 + Stage: Stage-5 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -516,87 +560,6 @@ STAGE PLANS: Table: default.nzhang_part8_n0 Is Table Level Stats: false - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - Reduce Output Operator - bucketingVersion: 2 - key expressions: '2008-12-31' (type: string), _col1 (type: string) - null sort order: zz - numBuckets: -1 - sort order: ++ - Map-reduce partition columns: '2008-12-31' (type: string), _col1 (type: string) - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -mr-10004 - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Truncated Path -> Alias: -#### A masked pattern was here #### - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: '2008-12-31' (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), '2008-12-31' (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - PREHOOK: query: from srcpart insert overwrite table nzhang_part8_n0 partition (ds, hr) select key, value, ds, hr where ds <= '2008-04-08' insert overwrite table nzhang_part8_n0 partition(ds='2008-12-31', hr) select key, value, hr where ds > '2008-04-08' diff --git ql/src/test/results/clientpositive/load_dyn_part9.q.out ql/src/test/results/clientpositive/llap/load_dyn_part9.q.out similarity index 87% rename from ql/src/test/results/clientpositive/load_dyn_part9.q.out rename to ql/src/test/results/clientpositive/llap/load_dyn_part9.q.out index 39bb39c297..451a0a2ddf 100644 --- ql/src/test/results/clientpositive/load_dyn_part9.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part9.q.out @@ -50,65 +50,79 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds <= '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part9 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds <= '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part9 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 546000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 2496 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -123,7 +137,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part9 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: diff --git ql/src/test/results/clientpositive/load_static_ptn_into_bucketed_table.q.out ql/src/test/results/clientpositive/llap/load_static_ptn_into_bucketed_table.q.out similarity index 75% rename from ql/src/test/results/clientpositive/load_static_ptn_into_bucketed_table.q.out rename to ql/src/test/results/clientpositive/llap/load_static_ptn_into_bucketed_table.q.out index 10b6dd398b..a27fa74205 100644 --- ql/src/test/results/clientpositive/load_static_ptn_into_bucketed_table.q.out +++ ql/src/test/results/clientpositive/llap/load_static_ptn_into_bucketed_table.q.out @@ -16,41 +16,54 @@ POSTHOOK: Input: default@src_bucket_tbl__temp_table_for_load_data__ POSTHOOK: Output: default@src_bucket_tbl@ds=2008-04-08 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_bucket_tbl__temp_table_for_load_data__ - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: a - sort order: + - Map-reduce partition columns: _col0 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_bucket_tbl__temp_table_for_load_data__ + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: a + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string) + outputColumnNames: _col0, _col1 Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket_tbl + File Output Operator + compressed: false + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src_bucket_tbl + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -64,7 +77,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src_bucket_tbl - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: @@ -231,41 +244,54 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src_bucket_tbl__temp_table_for_load_data__ STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_bucket_tbl__temp_table_for_load_data__ - Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string), ds (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col2 (type: string), _bucket_number (type: string), _col0 (type: int) - null sort order: aaa - sort order: +++ - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._bucket_number (type: string) - outputColumnNames: _col0, _col1, _col2, _bucket_number - File Output Operator - compressed: false - Dp Sort State: PARTITION_BUCKET_SORTED - Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket_tbl + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_bucket_tbl__temp_table_for_load_data__ + Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: string), ds (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string), _bucket_number (type: string), _col0 (type: int) + null sort order: aaa + sort order: +++ + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._bucket_number (type: string) + outputColumnNames: _col0, _col1, _col2, _bucket_number + File Output Operator + compressed: false + Dp Sort State: PARTITION_BUCKET_SORTED + Statistics: Num rows: 35 Data size: 13020 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src_bucket_tbl + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -280,7 +306,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src_bucket_tbl - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: @@ -447,41 +473,54 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src_bucket_tbl__temp_table_for_load_data__ STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_bucket_tbl__temp_table_for_load_data__ - Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string), ds (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col2 (type: string), _bucket_number (type: string), _col0 (type: int) - null sort order: aaa - sort order: +++ - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._bucket_number (type: string) - outputColumnNames: _col0, _col1, _col2, _bucket_number - File Output Operator - compressed: false - Dp Sort State: PARTITION_BUCKET_SORTED - Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket_tbl + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_bucket_tbl__temp_table_for_load_data__ + Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: string), ds (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string), _bucket_number (type: string), _col0 (type: int) + null sort order: aaa + sort order: +++ + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._bucket_number (type: string) + outputColumnNames: _col0, _col1, _col2, _bucket_number + File Output Operator + compressed: false + Dp Sort State: PARTITION_BUCKET_SORTED + Statistics: Num rows: 68 Data size: 24552 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src_bucket_tbl + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -496,7 +535,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src_bucket_tbl - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: @@ -664,41 +703,54 @@ POSTHOOK: Input: default@src_bucket_tbl__temp_table_for_load_data__ POSTHOOK: Output: default@src_bucket_tbl@hr=30/ds=2010-05-07 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src_bucket_tbl__temp_table_for_load_data__ - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: a - sort order: + - Map-reduce partition columns: _col0 (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src_bucket_tbl__temp_table_for_load_data__ + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: a + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string) + outputColumnNames: _col0, _col1 Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: int), VALUE._col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src_bucket_tbl + File Output Operator + compressed: false + Statistics: Num rows: 47 Data size: 8648 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src_bucket_tbl + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -713,7 +765,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src_bucket_tbl - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: diff --git ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out new file mode 100644 index 0000000000..f39ba92a88 --- /dev/null +++ ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out @@ -0,0 +1,1368 @@ +PREHOOK: query: EXPLAIN EXTENDED + FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key AND b.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED + FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key AND b.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` < 20 AND `key` > 15) AS `t0` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `key` > 15 AND `key` < 20 AND `ds` = '2008-04-08') AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [a] + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [b] + /srcpart/ds=2008-04-08/hr=12 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key AND b.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key AND b.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +17 val_17 17 val_17 +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +19 val_19 19 val_19 +PREHOOK: query: EXPLAIN EXTENDED + FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key AND a.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED + FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key AND a.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `key` < 20 AND `key` > 15 AND `ds` = '2008-04-08') AS `t0` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` > 15 AND `key` < 20) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [a] + /srcpart/ds=2008-04-08/hr=12 [a] + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key AND a.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key AND a.ds = '2008-04-08') + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +17 val_17 17 val_17 +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +19 val_19 19 val_19 +PREHOOK: query: EXPLAIN EXTENDED + FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED + FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` < 20 AND `key` > 15) AS `t0` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `key` > 15 AND `key` < 20 AND `ds` = '2008-04-08') AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [a] + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [b] + /srcpart/ds=2008-04-08/hr=12 [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: FROM + src a + LEFT OUTER JOIN + srcpart b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +17 val_17 17 val_17 +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +19 val_19 19 val_19 +PREHOOK: query: EXPLAIN EXTENDED + FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: EXPLAIN EXTENDED + FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +OPTIMIZED SQL: SELECT * +FROM (SELECT `key`, `value` +FROM `default`.`srcpart` +WHERE `key` < 20 AND `key` > 15 AND `ds` = '2008-04-08') AS `t0` +INNER JOIN (SELECT `key`, `value` +FROM `default`.`src` +WHERE `key` > 15 AND `key` < 20) AS `t2` ON `t0`.`key` = `t2`.`key` +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE + tag: 0 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [a] + /srcpart/ds=2008-04-08/hr=12 [a] + Map 3 + Map Operator Tree: + TableScan + alias: b + filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Filter Operator + isSamplingPred: false + predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE + tag: 1 + value expressions: _col1 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.src + numFiles 1 + numRows 500 + rawDataSize 5312 + serialization.ddl struct src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.src + name: default.src + Truncated Path -> Alias: + /src [b] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Position of Big Table: 0 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2,_col3 + columns.types string:string:string:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' +PREHOOK: type: QUERY +PREHOOK: Input: default@src +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: FROM + srcpart a + LEFT OUTER JOIN + src b + ON (a.key = b.key) + SELECT a.key, a.value, b.key, b.value + WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +#### A masked pattern was here #### +17 val_17 17 val_17 +17 val_17 17 val_17 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +18 val_18 18 val_18 +19 val_19 19 val_19 +19 val_19 19 val_19 diff --git ql/src/test/results/clientpositive/macro.q.out ql/src/test/results/clientpositive/llap/macro.q.out similarity index 81% rename from ql/src/test/results/clientpositive/macro.q.out rename to ql/src/test/results/clientpositive/llap/macro.q.out index c126a14d67..0e07ab09d3 100644 --- ql/src/test/results/clientpositive/macro.q.out +++ ql/src/test/results/clientpositive/llap/macro.q.out @@ -31,14 +31,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 0.8807970779778823D (type: double) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: EXPLAIN EXTENDED SELECT SIGMOID(2) FROM src LIMIT 1 @@ -62,15 +59,12 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE GatherStats: false Select Operator expressions: 0.8807970779778823D (type: double) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 4000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: DROP TEMPORARY MACRO SIGMOID @@ -112,14 +106,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 2 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: EXPLAIN EXTENDED SELECT FIXED_NUMBER() + 1 FROM src LIMIT 1 @@ -143,15 +134,12 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE GatherStats: false Select Operator expressions: 2 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: DROP TEMPORARY MACRO FIXED_NUMBER @@ -220,14 +208,11 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator expressions: 10 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: EXPLAIN EXTENDED SELECT SIMPLE_ADD(1, 9) FROM src LIMIT 1 @@ -251,15 +236,12 @@ STAGE PLANS: Processor Tree: TableScan alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE GatherStats: false Select Operator expressions: 10 (type: int) outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE ListSink PREHOOK: query: DROP TEMPORARY MACRO SIMPLE_ADD diff --git ql/src/test/results/clientpositive/mapreduce5.q.out ql/src/test/results/clientpositive/llap/mapreduce5.q.out similarity index 70% rename from ql/src/test/results/clientpositive/mapreduce5.q.out rename to ql/src/test/results/clientpositive/llap/mapreduce5.q.out index e4b6a528b8..4c014c3223 100644 --- ql/src/test/results/clientpositive/mapreduce5.q.out +++ ql/src/test/results/clientpositive/llap/mapreduce5.q.out @@ -26,58 +26,85 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@dest1_n133 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: int), _col2 (type: int) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col3 (type: string), _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col3 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(VALUE._col0) (type: int), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: int), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n133 - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: string) - outputColumnNames: key, ten, one, value - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(ten, 'hll'), compute_stats(one, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int), _col2 (type: int) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col3 (type: string), _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: UDFToInteger(VALUE._col0) (type: int), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: int), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n133 + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: string) + outputColumnNames: key, ten, one, value + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(ten, 'hll'), compute_stats(one, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -89,7 +116,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n133 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -97,30 +124,6 @@ STAGE PLANS: Column Types: int, int, int, string Table: default.dest1_n133 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM src INSERT OVERWRITE TABLE dest1_n133 SELECT src.key as c1, CAST(src.key / 10 AS INT) as c2, CAST(src.key % 10 AS INT) as c3, src.value as c4 diff --git ql/src/test/results/clientpositive/mapreduce6.q.out ql/src/test/results/clientpositive/llap/mapreduce6.q.out similarity index 70% rename from ql/src/test/results/clientpositive/mapreduce6.q.out rename to ql/src/test/results/clientpositive/llap/mapreduce6.q.out index d7ca82b48b..cf3e69b2e0 100644 --- ql/src/test/results/clientpositive/mapreduce6.q.out +++ ql/src/test/results/clientpositive/llap/mapreduce6.q.out @@ -26,58 +26,85 @@ POSTHOOK: Input: default@src POSTHOOK: Output: default@dest1_n169 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0, Stage-3 - Stage-3 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: int), _col2 (type: int) - null sort order: zz - sort order: -+ - Map-reduce partition columns: _col3 (type: string), _col0 (type: string) - Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col3 (type: string) - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: UDFToInteger(VALUE._col0) (type: int), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: int), VALUE._col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.dest1_n169 - Select Operator - expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: string) - outputColumnNames: key, ten, one, value - Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(ten, 'hll'), compute_stats(one, 'hll'), compute_stats(value, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Reducer 2 (CUSTOM_SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToInteger((key / 10)) (type: int), UDFToInteger((key % 10)) (type: int), value (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: int), _col2 (type: int) + null sort order: zz + sort order: -+ + Map-reduce partition columns: _col3 (type: string), _col0 (type: string) + Statistics: Num rows: 500 Data size: 93000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: UDFToInteger(VALUE._col0) (type: int), KEY.reducesinkkey0 (type: int), KEY.reducesinkkey1 (type: int), VALUE._col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.dest1_n169 + Select Operator + expressions: _col0 (type: int), _col1 (type: int), _col2 (type: int), _col3 (type: string) + outputColumnNames: key, ten, one, value + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(ten, 'hll'), compute_stats(one, 'hll'), compute_stats(value, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -89,7 +116,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n169 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -97,30 +124,6 @@ STAGE PLANS: Column Types: int, int, int, string Table: default.dest1_n169 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 1712 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct), _col1 (type: struct), _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1), compute_stats(VALUE._col2), compute_stats(VALUE._col3) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1760 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: FROM src INSERT OVERWRITE TABLE dest1_n169 SELECT src.key, CAST(src.key / 10 AS INT) as c2, CAST(src.key % 10 AS INT) as c3, src.value diff --git ql/src/test/results/clientpositive/llap/masking_1.q.out ql/src/test/results/clientpositive/llap/masking_1.q.out new file mode 100644 index 0000000000..ef42983d7a --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_1.q.out @@ -0,0 +1,432 @@ +PREHOOK: query: create table masking_test_n8 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n8 +POSTHOOK: query: create table masking_test_n8 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n8 +POSTHOOK: Lineage: masking_test_n8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n8.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select * from masking_test_n8 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n8 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n8 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n8 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +0 0_lav +0 0_lav +0 0_lav +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select key from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select key from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: select key from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select key from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +2 +4 +8 +PREHOOK: query: explain select value from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select value from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: select value from masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select value from masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +2_lav +4_lav +8_lav +PREHOOK: query: explain select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and ((key % 2) = 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), reverse(value) (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 5 Data size: 980 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: double) + Statistics: Num rows: 5 Data size: 980 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col4 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col4 (type: double) + Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col2 (type: double) + 1 _col4 (type: double) + outputColumnNames: _col0, _col1, _col3, _col4, _col5, _col6 + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +2 2_lav 2 val_2 2008-04-08 11 +2 2_lav 2 val_2 2008-04-08 12 +2 2_lav 2 val_2 2008-04-09 11 +2 2_lav 2 val_2 2008-04-09 12 +4 4_lav 4 val_4 2008-04-08 11 +4 4_lav 4 val_4 2008-04-08 12 +4 4_lav 4 val_4 2008-04-09 11 +4 4_lav 4 val_4 2008-04-09 12 +8 8_lav 8 val_8 2008-04-08 11 +8 8_lav 8 val_8 2008-04-08 12 +8 8_lav 8 val_8 2008-04-09 11 +8 8_lav 8 val_8 2008-04-09 12 +PREHOOK: query: explain select * from default.masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from default.masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from default.masking_test_n8 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select * from default.masking_test_n8 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n8 where masking_test_n8.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n8 where masking_test_n8.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n8 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n8 where masking_test_n8.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n8 where masking_test_n8.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n8 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) + Filter Operator + predicate: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) + Select Operator + expressions: key (type: string), upper(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + diff --git ql/src/test/results/clientpositive/llap/masking_10.q.out ql/src/test/results/clientpositive/llap/masking_10.q.out new file mode 100644 index 0000000000..39442c59cb --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_10.q.out @@ -0,0 +1,261 @@ +PREHOOK: query: drop table masking_test +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table masking_test +POSTHOOK: type: DROPTABLE +PREHOOK: query: create temporary table masking_test as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test +POSTHOOK: query: create temporary table masking_test as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test +PREHOOK: query: explain select 2017 as a, value from masking_test group by 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test +#### A masked pattern was here #### +POSTHOOK: query: explain select 2017 as a, value from masking_test group by 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: 2017 (type: int), _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select 2017 as a, value from masking_test group by 1, 2 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test +#### A masked pattern was here #### +POSTHOOK: query: select 2017 as a, value from masking_test group by 1, 2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test +#### A masked pattern was here #### +2017 2_lav +2017 4_lav +2017 0_lav +2017 8_lav +Warning: Shuffle Join MERGEJOIN[54][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +PREHOOK: query: explain +select * from + masking_test alias01 + left join + ( + select 2017 as a, value from masking_test group by 1, 2 + ) alias02 + on alias01.key = alias02.a + left join + masking_test alias03 +on alias01.key = alias03.key +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from + masking_test alias01 + left join + ( + select 2017 as a, value from masking_test group by 1, 2 + ) alias02 + on alias01.key = alias02.a + left join + masking_test alias03 +on alias01.key = alias03.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE), Reducer 4 (CUSTOM_SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test + filterExpr: (((key < 10) and ((key % 2) = 0)) or ((key < 10) and ((key % 2) = 0))) (type: boolean) + Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string) + Top N Key Operator + sort order: + + keys: reverse(value) (type: string) + null sort order: z + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + top n: 0 + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 + 1 + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 275 Data size: 49218 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 275 Data size: 49218 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE + Limit + Number of rows: 0 + Statistics: Num rows: 0 Data size: 0 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: 2017 (type: int), _col0 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE + value expressions: _col0 (type: int), _col1 (type: string) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap/masking_12.q.out ql/src/test/results/clientpositive/llap/masking_12.q.out new file mode 100644 index 0000000000..040e52cb66 --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_12.q.out @@ -0,0 +1,482 @@ +PREHOOK: query: create table `masking_test_n5` as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n5 +POSTHOOK: query: create table `masking_test_n5` as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n5 +POSTHOOK: Lineage: masking_test_n5.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n5.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: create view `v0` as select * from `masking_test_n5` +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Output: database:default +PREHOOK: Output: default@v0 +POSTHOOK: query: create view `v0` as select * from `masking_test_n5` +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@v0 +POSTHOOK: Lineage: v0.key SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:key, type:int, comment:null), ] +POSTHOOK: Lineage: v0.value SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:value, type:string, comment:null), ] +PREHOOK: query: explain +select * from `v0` +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from `v0` +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n5 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from `v0` +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: select * from `v0` +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +0 0_lav +4 4_lav +8 8_lav +0 0_lav +0 0_lav +2 2_lav +PREHOOK: query: create table `masking_test_subq_n1` as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_subq_n1 +POSTHOOK: query: create table `masking_test_subq_n1` as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_subq_n1 +POSTHOOK: Lineage: masking_test_subq_n1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_subq_n1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: create view `v1_n9` as select * from `masking_test_subq_n1` +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@masking_test_subq_n1 +PREHOOK: Output: database:default +PREHOOK: Output: default@v1_n9 +POSTHOOK: query: create view `v1_n9` as select * from `masking_test_subq_n1` +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@masking_test_subq_n1 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@v1_n9 +POSTHOOK: Lineage: v1_n9.key SIMPLE [(masking_test_subq_n1)masking_test_subq_n1.FieldSchema(name:key, type:int, comment:null), ] +POSTHOOK: Lineage: v1_n9.value SIMPLE [(masking_test_subq_n1)masking_test_subq_n1.FieldSchema(name:value, type:string, comment:null), ] +PREHOOK: query: explain +select * from `v1_n9` +limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n1 +PREHOOK: Input: default@src +PREHOOK: Input: default@v1_n9 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from `v1_n9` +limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n1 +POSTHOOK: Input: default@src +POSTHOOK: Input: default@v1_n9 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n1 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n1 + filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 20 + Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: 20 + Processor Tree: + ListSink + +PREHOOK: query: select * from `v1_n9` +limit 20 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n1 +PREHOOK: Input: default@src +PREHOOK: Input: default@v1_n9 +#### A masked pattern was here #### +POSTHOOK: query: select * from `v1_n9` +limit 20 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n1 +POSTHOOK: Input: default@src +POSTHOOK: Input: default@v1_n9 +#### A masked pattern was here #### +0 val_0 +0 val_0 +0 val_0 +2 val_2 +4 val_4 +5 val_5 +5 val_5 +5 val_5 +8 val_8 +9 val_9 +10 val_10 +11 val_11 +12 val_12 +12 val_12 +15 val_15 +15 val_15 +17 val_17 +18 val_18 +18 val_18 +19 val_19 +PREHOOK: query: create view `masking_test_view` as select key from `v0` +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@v0 +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_view +POSTHOOK: query: create view `masking_test_view` as select key from `v0` +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@v0 +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_view +POSTHOOK: Lineage: masking_test_view.key SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:key, type:int, comment:null), ] +PREHOOK: query: explain +select key from `masking_test_view` +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@masking_test_view +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: explain +select key from `masking_test_view` +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@masking_test_view +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n5 + filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: select key from `masking_test_view` +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@masking_test_view +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: select key from `masking_test_view` +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@masking_test_view +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +4 +PREHOOK: query: explain +select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@masking_test_view +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: explain +select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@masking_test_view +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n5 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: masking_test_n5 + filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0) and ((UDFToInteger((UDFToDouble(key) / 2.0D)) % 2) = 0) and (UDFToInteger((UDFToDouble(key) / 2.0D)) < 10)) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and (key > 6) and ((key % 2) = 0) and ((UDFToInteger((UDFToDouble(key) / 2.0D)) % 2) = 0) and (UDFToInteger((UDFToDouble(key) / 2.0D)) < 10)) (type: boolean) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col1 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n5 +PREHOOK: Input: default@masking_test_view +PREHOOK: Input: default@v0 +#### A masked pattern was here #### +POSTHOOK: query: select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n5 +POSTHOOK: Input: default@masking_test_view +POSTHOOK: Input: default@v0 +#### A masked pattern was here #### +4_lav diff --git ql/src/test/results/clientpositive/masking_13.q.out ql/src/test/results/clientpositive/llap/masking_13.q.out similarity index 69% rename from ql/src/test/results/clientpositive/masking_13.q.out rename to ql/src/test/results/clientpositive/llap/masking_13.q.out index 672d65fd31..5fdc36b9f0 100644 --- ql/src/test/results/clientpositive/masking_13.q.out +++ ql/src/test/results/clientpositive/llap/masking_13.q.out @@ -19,38 +19,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select * from masking_test PREHOOK: type: QUERY @@ -118,40 +102,24 @@ POSTHOOK: Input: default@masking_test POSTHOOK: Input: default@masking_test_view #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test + filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: select key from `masking_test_view` PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/masking_1_newdb.q.out ql/src/test/results/clientpositive/llap/masking_1_newdb.q.out similarity index 51% rename from ql/src/test/results/clientpositive/masking_1_newdb.q.out rename to ql/src/test/results/clientpositive/llap/masking_1_newdb.q.out index 293da26b7a..864bb50d61 100644 --- ql/src/test/results/clientpositive/masking_1_newdb.q.out +++ ql/src/test/results/clientpositive/llap/masking_1_newdb.q.out @@ -37,38 +37,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: newdb@masking_test_n12 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n12 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n12 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select * from newdb.masking_test_n12 PREHOOK: type: QUERY @@ -93,38 +77,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: newdb@masking_test_n12 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n12 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n12 + filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink PREHOOK: query: select * from newdb.masking_test_n12 where key > 0 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/masking_2.q.out ql/src/test/results/clientpositive/llap/masking_2.q.out new file mode 100644 index 0000000000..2042e732e6 --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_2.q.out @@ -0,0 +1,391 @@ +PREHOOK: query: create view masking_test_n1 as select cast(key as int) as key, value from src +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n1 +POSTHOOK: query: create view masking_test_n1 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n1 +POSTHOOK: Lineage: masking_test_n1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select * from masking_test_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n1 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 0_lav +0 0_lav +0 0_lav +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n1 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n1 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n1 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n1 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from src a join masking_test_n1 b on a.key = b.value where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from src a join masking_test_n1 b on a.key = b.value where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and reverse(value) is not null) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and reverse(value) is not null) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select a.*, b.key from masking_test_n1 a join masking_test_n1 b on a.key = b.value where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select a.*, b.key from masking_test_n1 a join masking_test_n1 b on a.key = b.value where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) + Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col0) (type: double) + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and UDFToDouble(reverse(value)) is not null) (type: boolean) + properties: + insideView TRUE + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and UDFToDouble(reverse(value)) is not null) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col1) (type: double) + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double) + 1 UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * from masking_test_n1 a union select b.* from masking_test_n1 b where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n1 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n1 a union select b.* from masking_test_n1 b where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n1 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: int), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: int), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 2 + Vertex: Union 2 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/masking_3.q.out ql/src/test/results/clientpositive/llap/masking_3.q.out similarity index 78% rename from ql/src/test/results/clientpositive/masking_3.q.out rename to ql/src/test/results/clientpositive/llap/masking_3.q.out index 0a12994a03..a359d3c6cb 100644 --- ql/src/test/results/clientpositive/masking_3.q.out +++ ql/src/test/results/clientpositive/llap/masking_3.q.out @@ -21,151 +21,150 @@ POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) Stage: Stage-0 Fetch Operator @@ -186,87 +185,7 @@ POSTHOOK: Input: default@src 0 val_0 0 val_0 0 val_0 -2 val_2 -4 val_4 -5 val_5 -5 val_5 -5 val_5 -8 val_8 -9 val_9 10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -80 val_80 -82 val_82 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -87 val_87 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -98 val_98 -98 val_98 100 val_100 100 val_100 103 val_103 @@ -274,6 +193,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -284,6 +204,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -311,6 +233,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -335,6 +259,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -348,6 +273,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -356,6 +283,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -372,6 +300,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -422,6 +352,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -436,12 +368,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -455,6 +390,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -475,6 +411,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -506,6 +443,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -515,6 +453,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -526,6 +465,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -541,6 +483,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -566,6 +510,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -586,6 +531,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -596,12 +542,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -617,6 +566,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -651,6 +601,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -683,217 +634,9 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 -PREHOOK: query: explain select * from masking_test_subq_n3 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n3 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_subq_n3 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n3 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_subq_n3 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n3 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_subq_n3 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n3 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -2 val_2 -4 val_4 5 val_5 5 val_5 5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 51 val_51 51 val_51 53 val_53 @@ -917,6 +660,7 @@ POSTHOOK: Input: default@src 76 val_76 77 val_77 78 val_78 +8 val_8 80 val_80 82 val_82 83 val_83 @@ -926,6 +670,7 @@ POSTHOOK: Input: default@src 85 val_85 86 val_86 87 val_87 +9 val_9 90 val_90 90 val_90 90 val_90 @@ -937,50 +682,228 @@ POSTHOOK: Input: default@src 97 val_97 98 val_98 98 val_98 -100 val_100 -100 val_100 -103 val_103 -103 val_103 -104 val_104 -104 val_104 -105 val_105 -111 val_111 -113 val_113 -113 val_113 -114 val_114 -116 val_116 -118 val_118 -118 val_118 -119 val_119 -119 val_119 -119 val_119 -120 val_120 -120 val_120 -125 val_125 -125 val_125 -126 val_126 -128 val_128 -128 val_128 -128 val_128 -129 val_129 -129 val_129 -131 val_131 -133 val_133 -134 val_134 -134 val_134 -136 val_136 -137 val_137 -137 val_137 -138 val_138 -138 val_138 -138 val_138 -138 val_138 -143 val_143 -145 val_145 -146 val_146 -146 val_146 -149 val_149 -149 val_149 +PREHOOK: query: explain select * from masking_test_subq_n3 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n3 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_subq_n3 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n3 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_subq_n3 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n3 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_subq_n3 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n3 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 +100 val_100 +100 val_100 +103 val_103 +103 val_103 +104 val_104 +104 val_104 +105 val_105 +11 val_11 +111 val_111 +113 val_113 +113 val_113 +114 val_114 +116 val_116 +118 val_118 +118 val_118 +119 val_119 +119 val_119 +119 val_119 +12 val_12 +12 val_12 +120 val_120 +120 val_120 +125 val_125 +125 val_125 +126 val_126 +128 val_128 +128 val_128 +128 val_128 +129 val_129 +129 val_129 +131 val_131 +133 val_133 +134 val_134 +134 val_134 +136 val_136 +137 val_137 +137 val_137 +138 val_138 +138 val_138 +138 val_138 +138 val_138 +143 val_143 +145 val_145 +146 val_146 +146 val_146 +149 val_149 +149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -1005,6 +928,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -1018,6 +942,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -1026,6 +952,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -1042,6 +969,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -1092,6 +1021,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -1106,12 +1037,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -1125,6 +1059,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -1145,6 +1080,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -1176,6 +1112,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -1185,6 +1122,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -1196,6 +1134,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -1211,6 +1152,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -1236,6 +1179,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -1256,6 +1200,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -1266,12 +1211,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -1287,6 +1235,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -1321,6 +1270,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -1353,6 +1303,54 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +80 val_80 +82 val_82 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +87 val_87 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +98 val_98 +98 val_98 PREHOOK: query: explain select key from masking_test_subq_n3 where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq_n3 @@ -1364,150 +1362,138 @@ POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) + Reducer 4 <- Map 5 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 325 Data size: 1300 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 1300 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 325 Data size: 1300 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 1300 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE Stage: Stage-0 Fetch Operator @@ -1525,131 +1511,56 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### -2 -4 -5 -5 -5 -8 -9 10 +100 +100 +103 +103 +104 +104 +105 11 +111 +113 +113 +114 +116 +118 +118 +119 +119 +119 12 12 +120 +120 +125 +125 +126 +128 +128 +128 +129 +129 +131 +133 +134 +134 +136 +137 +137 +138 +138 +138 +138 +143 +145 +146 +146 +149 +149 15 15 -17 -18 -18 -19 -20 -24 -24 -26 -26 -27 -28 -30 -33 -34 -35 -35 -35 -37 -37 -41 -42 -42 -43 -44 -47 -51 -51 -53 -54 -57 -58 -58 -64 -65 -66 -67 -67 -69 -70 -70 -70 -72 -72 -74 -76 -76 -77 -78 -80 -82 -83 -83 -84 -84 -85 -86 -87 -90 -90 -90 -92 -95 -95 -96 -97 -97 -98 -98 -100 -100 -103 -103 -104 -104 -105 -111 -113 -113 -114 -116 -118 -118 -119 -119 -119 -120 -120 -125 -125 -126 -128 -128 -128 -129 -129 -131 -133 -134 -134 -136 -137 -137 -138 -138 -138 -138 -143 -145 -146 -146 -149 -149 150 152 152 @@ -1674,6 +1585,7 @@ POSTHOOK: Input: default@src 169 169 169 +17 170 172 172 @@ -1687,6 +1599,8 @@ POSTHOOK: Input: default@src 178 179 179 +18 +18 180 181 183 @@ -1695,6 +1609,7 @@ POSTHOOK: Input: default@src 187 187 189 +19 190 191 191 @@ -1711,6 +1626,8 @@ POSTHOOK: Input: default@src 199 199 199 +2 +20 200 200 201 @@ -1761,6 +1678,8 @@ POSTHOOK: Input: default@src 238 239 239 +24 +24 241 242 242 @@ -1775,12 +1694,15 @@ POSTHOOK: Input: default@src 256 257 258 +26 +26 260 262 263 265 265 266 +27 272 272 273 @@ -1794,6 +1716,7 @@ POSTHOOK: Input: default@src 277 278 278 +28 280 280 281 @@ -1814,6 +1737,7 @@ POSTHOOK: Input: default@src 298 298 298 +30 302 305 306 @@ -1845,6 +1769,7 @@ POSTHOOK: Input: default@src 327 327 327 +33 331 331 332 @@ -1854,6 +1779,7 @@ POSTHOOK: Input: default@src 336 338 339 +34 341 342 342 @@ -1865,6 +1791,9 @@ POSTHOOK: Input: default@src 348 348 348 +35 +35 +35 351 353 353 @@ -1880,6 +1809,8 @@ POSTHOOK: Input: default@src 369 369 369 +37 +37 373 374 375 @@ -1905,6 +1836,7 @@ POSTHOOK: Input: default@src 397 399 399 +4 400 401 401 @@ -1925,6 +1857,7 @@ POSTHOOK: Input: default@src 409 409 409 +41 411 413 413 @@ -1935,12 +1868,15 @@ POSTHOOK: Input: default@src 417 418 419 +42 +42 421 424 424 427 429 429 +43 430 430 430 @@ -1956,6 +1892,7 @@ POSTHOOK: Input: default@src 438 439 439 +44 443 444 446 @@ -1990,6 +1927,7 @@ POSTHOOK: Input: default@src 469 469 469 +47 470 472 475 @@ -2022,6 +1960,54 @@ POSTHOOK: Input: default@src 498 498 498 +5 +5 +5 +51 +51 +53 +54 +57 +58 +58 +64 +65 +66 +67 +67 +69 +70 +70 +70 +72 +72 +74 +76 +76 +77 +78 +8 +80 +82 +83 +83 +84 +84 +85 +86 +87 +9 +90 +90 +90 +92 +95 +95 +96 +97 +97 +98 +98 PREHOOK: query: explain select value from masking_test_subq_n3 where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq_n3 @@ -2033,155 +2019,154 @@ POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col1 + Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col1 - Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 29575 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) Stage: Stage-0 Fetch Operator @@ -2199,131 +2184,56 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### -val_2 -val_4 -val_5 -val_5 -val_5 -val_8 -val_9 val_10 +val_100 +val_100 +val_103 +val_103 +val_104 +val_104 +val_105 val_11 +val_111 +val_113 +val_113 +val_114 +val_116 +val_118 +val_118 +val_119 +val_119 +val_119 val_12 val_12 +val_120 +val_120 +val_125 +val_125 +val_126 +val_128 +val_128 +val_128 +val_129 +val_129 +val_131 +val_133 +val_134 +val_134 +val_136 +val_137 +val_137 +val_138 +val_138 +val_138 +val_138 +val_143 +val_145 +val_146 +val_146 +val_149 +val_149 val_15 val_15 -val_17 -val_18 -val_18 -val_19 -val_20 -val_24 -val_24 -val_26 -val_26 -val_27 -val_28 -val_30 -val_33 -val_34 -val_35 -val_35 -val_35 -val_37 -val_37 -val_41 -val_42 -val_42 -val_43 -val_44 -val_47 -val_51 -val_51 -val_53 -val_54 -val_57 -val_58 -val_58 -val_64 -val_65 -val_66 -val_67 -val_67 -val_69 -val_70 -val_70 -val_70 -val_72 -val_72 -val_74 -val_76 -val_76 -val_77 -val_78 -val_80 -val_82 -val_83 -val_83 -val_84 -val_84 -val_85 -val_86 -val_87 -val_90 -val_90 -val_90 -val_92 -val_95 -val_95 -val_96 -val_97 -val_97 -val_98 -val_98 -val_100 -val_100 -val_103 -val_103 -val_104 -val_104 -val_105 -val_111 -val_113 -val_113 -val_114 -val_116 -val_118 -val_118 -val_119 -val_119 -val_119 -val_120 -val_120 -val_125 -val_125 -val_126 -val_128 -val_128 -val_128 -val_129 -val_129 -val_131 -val_133 -val_134 -val_134 -val_136 -val_137 -val_137 -val_138 -val_138 -val_138 -val_138 -val_143 -val_145 -val_146 -val_146 -val_149 -val_149 val_150 val_152 val_152 @@ -2348,6 +2258,7 @@ val_169 val_169 val_169 val_169 +val_17 val_170 val_172 val_172 @@ -2361,6 +2272,8 @@ val_177 val_178 val_179 val_179 +val_18 +val_18 val_180 val_181 val_183 @@ -2369,6 +2282,7 @@ val_187 val_187 val_187 val_189 +val_19 val_190 val_191 val_191 @@ -2385,6 +2299,8 @@ val_197 val_199 val_199 val_199 +val_2 +val_20 val_200 val_200 val_201 @@ -2435,6 +2351,8 @@ val_238 val_238 val_239 val_239 +val_24 +val_24 val_241 val_242 val_242 @@ -2449,12 +2367,15 @@ val_256 val_256 val_257 val_258 +val_26 +val_26 val_260 val_262 val_263 val_265 val_265 val_266 +val_27 val_272 val_272 val_273 @@ -2468,6 +2389,7 @@ val_277 val_277 val_278 val_278 +val_28 val_280 val_280 val_281 @@ -2488,6 +2410,7 @@ val_296 val_298 val_298 val_298 +val_30 val_302 val_305 val_306 @@ -2519,6 +2442,7 @@ val_325 val_327 val_327 val_327 +val_33 val_331 val_331 val_332 @@ -2528,6 +2452,7 @@ val_335 val_336 val_338 val_339 +val_34 val_341 val_342 val_342 @@ -2539,6 +2464,9 @@ val_348 val_348 val_348 val_348 +val_35 +val_35 +val_35 val_351 val_353 val_353 @@ -2554,6 +2482,8 @@ val_368 val_369 val_369 val_369 +val_37 +val_37 val_373 val_374 val_375 @@ -2579,6 +2509,7 @@ val_397 val_397 val_399 val_399 +val_4 val_400 val_401 val_401 @@ -2599,6 +2530,7 @@ val_407 val_409 val_409 val_409 +val_41 val_411 val_413 val_413 @@ -2609,12 +2541,15 @@ val_417 val_417 val_418 val_419 +val_42 +val_42 val_421 val_424 val_424 val_427 val_429 val_429 +val_43 val_430 val_430 val_430 @@ -2630,6 +2565,7 @@ val_438 val_438 val_439 val_439 +val_44 val_443 val_444 val_446 @@ -2664,6 +2600,7 @@ val_469 val_469 val_469 val_469 +val_47 val_470 val_472 val_475 @@ -2696,6 +2633,54 @@ val_497 val_498 val_498 val_498 +val_5 +val_5 +val_5 +val_51 +val_51 +val_53 +val_54 +val_57 +val_58 +val_58 +val_64 +val_65 +val_66 +val_67 +val_67 +val_69 +val_70 +val_70 +val_70 +val_72 +val_72 +val_74 +val_76 +val_76 +val_77 +val_78 +val_8 +val_80 +val_82 +val_83 +val_83 +val_84 +val_84 +val_85 +val_86 +val_87 +val_9 +val_90 +val_90 +val_90 +val_92 +val_95 +val_95 +val_96 +val_97 +val_97 +val_98 +val_98 PREHOOK: query: explain select * from masking_test_subq_n3 join srcpart on (masking_test_subq_n3.key = srcpart.key) PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq_n3 @@ -2717,205 +2702,199 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-4 is a root stage - Stage-3 depends on stages: Stage-4 - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE), Reducer 8 (SIMPLE_EDGE) + Reducer 8 <- Map 7 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col4 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col4 (type: double) + Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col4 (type: double) + 1 _col2 (type: double) + outputColumnNames: _col0, _col1, _col2, _col3, _col5, _col6 + Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col5 (type: int), _col6 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: string), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 325 Data size: 33475 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col2 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col2 (type: double) + Statistics: Num rows: 325 Data size: 33475 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Reducer 6 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 8 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 325 Data size: 33475 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col4 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col4 (type: double) - Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - TableScan - Reduce Output Operator - key expressions: _col2 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: double) - Statistics: Num rows: 325 Data size: 33475 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col4 (type: double) - 1 _col2 (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col5, _col6 - Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col5 (type: int), _col6 (type: string), _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2056 Data size: 1317896 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) Stage: Stage-0 Fetch Operator @@ -2943,656 +2922,108 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 #### A masked pattern was here #### -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-08 11 0 val_0 0 val_0 2008-04-08 11 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 11 0 val_0 0 val_0 2008-04-08 11 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-08 11 +0 val_0 0 val_0 2008-04-08 11 +0 val_0 0 val_0 2008-04-08 11 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-09 12 -2 val_2 2 val_2 2008-04-09 11 -2 val_2 2 val_2 2008-04-08 11 -2 val_2 2 val_2 2008-04-09 12 -2 val_2 2 val_2 2008-04-08 12 -4 val_4 4 val_4 2008-04-08 12 -4 val_4 4 val_4 2008-04-09 12 -4 val_4 4 val_4 2008-04-08 11 -4 val_4 4 val_4 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-09 12 -8 val_8 8 val_8 2008-04-08 11 -8 val_8 8 val_8 2008-04-09 11 -8 val_8 8 val_8 2008-04-08 12 -8 val_8 8 val_8 2008-04-09 12 -9 val_9 9 val_9 2008-04-08 12 -9 val_9 9 val_9 2008-04-09 12 -9 val_9 9 val_9 2008-04-09 11 -9 val_9 9 val_9 2008-04-08 11 -10 val_10 10 val_10 2008-04-09 11 -10 val_10 10 val_10 2008-04-09 12 -10 val_10 10 val_10 2008-04-08 12 -10 val_10 10 val_10 2008-04-08 11 -11 val_11 11 val_11 2008-04-09 12 -11 val_11 11 val_11 2008-04-09 11 -11 val_11 11 val_11 2008-04-08 12 -11 val_11 11 val_11 2008-04-08 11 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-09 12 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 12 -17 val_17 17 val_17 2008-04-09 11 -17 val_17 17 val_17 2008-04-09 12 -17 val_17 17 val_17 2008-04-08 12 -17 val_17 17 val_17 2008-04-08 11 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-08 11 -19 val_19 19 val_19 2008-04-09 11 -19 val_19 19 val_19 2008-04-09 12 -19 val_19 19 val_19 2008-04-08 11 -19 val_19 19 val_19 2008-04-08 12 -20 val_20 20 val_20 2008-04-08 11 -20 val_20 20 val_20 2008-04-09 11 -20 val_20 20 val_20 2008-04-08 12 -20 val_20 20 val_20 2008-04-09 12 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-08 11 -27 val_27 27 val_27 2008-04-09 12 -27 val_27 27 val_27 2008-04-09 11 -27 val_27 27 val_27 2008-04-08 12 -27 val_27 27 val_27 2008-04-08 11 -28 val_28 28 val_28 2008-04-08 12 -28 val_28 28 val_28 2008-04-09 11 -28 val_28 28 val_28 2008-04-09 12 -28 val_28 28 val_28 2008-04-08 11 -30 val_30 30 val_30 2008-04-09 12 -30 val_30 30 val_30 2008-04-09 11 -30 val_30 30 val_30 2008-04-08 11 -30 val_30 30 val_30 2008-04-08 12 -33 val_33 33 val_33 2008-04-08 11 -33 val_33 33 val_33 2008-04-08 12 -33 val_33 33 val_33 2008-04-09 11 -33 val_33 33 val_33 2008-04-09 12 -34 val_34 34 val_34 2008-04-09 11 -34 val_34 34 val_34 2008-04-09 12 -34 val_34 34 val_34 2008-04-08 11 -34 val_34 34 val_34 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-08 11 -41 val_41 41 val_41 2008-04-08 11 -41 val_41 41 val_41 2008-04-09 11 -41 val_41 41 val_41 2008-04-09 12 -41 val_41 41 val_41 2008-04-08 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-08 11 -43 val_43 43 val_43 2008-04-09 12 -43 val_43 43 val_43 2008-04-08 12 -43 val_43 43 val_43 2008-04-09 11 -43 val_43 43 val_43 2008-04-08 11 -44 val_44 44 val_44 2008-04-09 11 -44 val_44 44 val_44 2008-04-08 11 -44 val_44 44 val_44 2008-04-09 12 -44 val_44 44 val_44 2008-04-08 12 -47 val_47 47 val_47 2008-04-08 11 -47 val_47 47 val_47 2008-04-08 12 -47 val_47 47 val_47 2008-04-09 12 -47 val_47 47 val_47 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 11 -53 val_53 53 val_53 2008-04-08 11 -53 val_53 53 val_53 2008-04-09 11 -53 val_53 53 val_53 2008-04-09 12 -53 val_53 53 val_53 2008-04-08 12 -54 val_54 54 val_54 2008-04-08 11 -54 val_54 54 val_54 2008-04-09 12 -54 val_54 54 val_54 2008-04-09 11 -54 val_54 54 val_54 2008-04-08 12 -57 val_57 57 val_57 2008-04-08 12 -57 val_57 57 val_57 2008-04-08 11 -57 val_57 57 val_57 2008-04-09 11 -57 val_57 57 val_57 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-08 12 -64 val_64 64 val_64 2008-04-09 12 -64 val_64 64 val_64 2008-04-08 11 -64 val_64 64 val_64 2008-04-08 12 -64 val_64 64 val_64 2008-04-09 11 -65 val_65 65 val_65 2008-04-09 11 -65 val_65 65 val_65 2008-04-09 12 -65 val_65 65 val_65 2008-04-08 11 -65 val_65 65 val_65 2008-04-08 12 -66 val_66 66 val_66 2008-04-09 12 -66 val_66 66 val_66 2008-04-09 11 -66 val_66 66 val_66 2008-04-08 11 -66 val_66 66 val_66 2008-04-08 12 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-08 12 -69 val_69 69 val_69 2008-04-09 12 -69 val_69 69 val_69 2008-04-08 11 -69 val_69 69 val_69 2008-04-08 12 -69 val_69 69 val_69 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 12 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-09 11 -74 val_74 74 val_74 2008-04-09 11 -74 val_74 74 val_74 2008-04-08 12 -74 val_74 74 val_74 2008-04-09 12 -74 val_74 74 val_74 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-09 11 -77 val_77 77 val_77 2008-04-08 12 -77 val_77 77 val_77 2008-04-09 12 -77 val_77 77 val_77 2008-04-09 11 -77 val_77 77 val_77 2008-04-08 11 -78 val_78 78 val_78 2008-04-09 11 -78 val_78 78 val_78 2008-04-08 11 -78 val_78 78 val_78 2008-04-09 12 -78 val_78 78 val_78 2008-04-08 12 -80 val_80 80 val_80 2008-04-09 11 -80 val_80 80 val_80 2008-04-08 11 -80 val_80 80 val_80 2008-04-09 12 -80 val_80 80 val_80 2008-04-08 12 -82 val_82 82 val_82 2008-04-08 12 -82 val_82 82 val_82 2008-04-09 11 -82 val_82 82 val_82 2008-04-08 11 -82 val_82 82 val_82 2008-04-09 12 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 11 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-08 12 -85 val_85 85 val_85 2008-04-09 12 -85 val_85 85 val_85 2008-04-09 11 -85 val_85 85 val_85 2008-04-08 11 -85 val_85 85 val_85 2008-04-08 12 -86 val_86 86 val_86 2008-04-08 12 -86 val_86 86 val_86 2008-04-09 11 -86 val_86 86 val_86 2008-04-08 11 -86 val_86 86 val_86 2008-04-09 12 -87 val_87 87 val_87 2008-04-09 12 -87 val_87 87 val_87 2008-04-08 11 -87 val_87 87 val_87 2008-04-08 12 -87 val_87 87 val_87 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 11 -92 val_92 92 val_92 2008-04-09 11 -92 val_92 92 val_92 2008-04-09 12 -92 val_92 92 val_92 2008-04-08 12 -92 val_92 92 val_92 2008-04-08 11 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-08 11 -96 val_96 96 val_96 2008-04-09 12 -96 val_96 96 val_96 2008-04-08 11 -96 val_96 96 val_96 2008-04-08 12 -96 val_96 96 val_96 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-09 12 -100 val_100 100 val_100 2008-04-09 11 -100 val_100 100 val_100 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +10 val_10 10 val_10 2008-04-08 11 +10 val_10 10 val_10 2008-04-08 12 +10 val_10 10 val_10 2008-04-09 11 +10 val_10 10 val_10 2008-04-09 12 +100 val_100 100 val_100 2008-04-08 11 +100 val_100 100 val_100 2008-04-08 11 100 val_100 100 val_100 2008-04-08 11 100 val_100 100 val_100 2008-04-08 11 -100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-09 12 100 val_100 100 val_100 2008-04-08 12 100 val_100 100 val_100 2008-04-08 12 100 val_100 100 val_100 2008-04-08 12 100 val_100 100 val_100 2008-04-08 12 -100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-08 11 -100 val_100 100 val_100 2008-04-08 11 100 val_100 100 val_100 2008-04-09 11 100 val_100 100 val_100 2008-04-09 11 +100 val_100 100 val_100 2008-04-09 11 +100 val_100 100 val_100 2008-04-09 11 +100 val_100 100 val_100 2008-04-09 12 +100 val_100 100 val_100 2008-04-09 12 +100 val_100 100 val_100 2008-04-09 12 +100 val_100 100 val_100 2008-04-09 12 103 val_103 103 val_103 2008-04-08 11 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 12 -103 val_103 103 val_103 2008-04-09 12 -103 val_103 103 val_103 2008-04-09 11 -103 val_103 103 val_103 2008-04-09 11 103 val_103 103 val_103 2008-04-08 11 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 11 -103 val_103 103 val_103 2008-04-09 11 103 val_103 103 val_103 2008-04-08 12 103 val_103 103 val_103 2008-04-08 12 103 val_103 103 val_103 2008-04-08 12 103 val_103 103 val_103 2008-04-08 12 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 12 +103 val_103 103 val_103 2008-04-09 12 103 val_103 103 val_103 2008-04-09 12 103 val_103 103 val_103 2008-04-09 12 104 val_104 104 val_104 2008-04-08 11 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-09 11 -104 val_104 104 val_104 2008-04-09 11 -104 val_104 104 val_104 2008-04-08 12 -104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 11 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-08 12 +104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-09 11 104 val_104 104 val_104 2008-04-09 11 -105 val_105 105 val_105 2008-04-09 12 -105 val_105 105 val_105 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 105 val_105 105 val_105 2008-04-08 11 105 val_105 105 val_105 2008-04-08 12 -111 val_111 111 val_111 2008-04-08 12 +105 val_105 105 val_105 2008-04-09 11 +105 val_105 105 val_105 2008-04-09 12 +11 val_11 11 val_11 2008-04-08 11 +11 val_11 11 val_11 2008-04-08 12 +11 val_11 11 val_11 2008-04-09 11 +11 val_11 11 val_11 2008-04-09 12 111 val_111 111 val_111 2008-04-08 11 -111 val_111 111 val_111 2008-04-09 12 +111 val_111 111 val_111 2008-04-08 12 111 val_111 111 val_111 2008-04-09 11 -113 val_113 113 val_113 2008-04-09 12 -113 val_113 113 val_113 2008-04-09 12 +111 val_111 111 val_111 2008-04-09 12 +113 val_113 113 val_113 2008-04-08 11 +113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-08 12 @@ -3603,26 +3034,20 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 113 val_113 113 val_113 2008-04-09 11 113 val_113 113 val_113 2008-04-09 11 113 val_113 113 val_113 2008-04-09 11 -113 val_113 113 val_113 2008-04-08 11 -113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-09 12 113 val_113 113 val_113 2008-04-09 12 +113 val_113 113 val_113 2008-04-09 12 +113 val_113 113 val_113 2008-04-09 12 +114 val_114 114 val_114 2008-04-08 11 +114 val_114 114 val_114 2008-04-08 12 114 val_114 114 val_114 2008-04-09 11 114 val_114 114 val_114 2008-04-09 12 -114 val_114 114 val_114 2008-04-08 12 -114 val_114 114 val_114 2008-04-08 11 +116 val_116 116 val_116 2008-04-08 11 116 val_116 116 val_116 2008-04-08 12 116 val_116 116 val_116 2008-04-09 11 -116 val_116 116 val_116 2008-04-08 11 116 val_116 116 val_116 2008-04-09 12 -118 val_118 118 val_118 2008-04-09 12 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 118 val_118 118 val_118 2008-04-08 11 -118 val_118 118 val_118 2008-04-09 11 -118 val_118 118 val_118 2008-04-09 11 -118 val_118 118 val_118 2008-04-09 12 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 118 val_118 118 val_118 2008-04-08 11 118 val_118 118 val_118 2008-04-08 12 @@ -3631,15 +3056,21 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 118 val_118 118 val_118 2008-04-08 12 118 val_118 118 val_118 2008-04-09 11 118 val_118 118 val_118 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 11 +118 val_118 118 val_118 2008-04-09 11 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 @@ -3649,9 +3080,9 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 11 @@ -3661,54 +3092,67 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-09 12 -120 val_120 120 val_120 2008-04-09 11 -120 val_120 120 val_120 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +120 val_120 120 val_120 2008-04-08 11 +120 val_120 120 val_120 2008-04-08 11 120 val_120 120 val_120 2008-04-08 11 120 val_120 120 val_120 2008-04-08 11 -120 val_120 120 val_120 2008-04-09 12 -120 val_120 120 val_120 2008-04-09 12 120 val_120 120 val_120 2008-04-08 12 120 val_120 120 val_120 2008-04-08 12 -120 val_120 120 val_120 2008-04-09 12 -120 val_120 120 val_120 2008-04-09 12 120 val_120 120 val_120 2008-04-08 12 120 val_120 120 val_120 2008-04-08 12 -120 val_120 120 val_120 2008-04-08 11 -120 val_120 120 val_120 2008-04-08 11 120 val_120 120 val_120 2008-04-09 11 120 val_120 120 val_120 2008-04-09 11 -125 val_125 125 val_125 2008-04-09 12 -125 val_125 125 val_125 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 11 +120 val_120 120 val_120 2008-04-09 11 +120 val_120 120 val_120 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 12 125 val_125 125 val_125 2008-04-08 11 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 11 -125 val_125 125 val_125 2008-04-09 11 125 val_125 125 val_125 2008-04-08 11 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 12 -125 val_125 125 val_125 2008-04-09 12 -125 val_125 125 val_125 2008-04-09 11 -125 val_125 125 val_125 2008-04-09 11 125 val_125 125 val_125 2008-04-08 12 125 val_125 125 val_125 2008-04-08 12 125 val_125 125 val_125 2008-04-08 12 125 val_125 125 val_125 2008-04-08 12 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 126 val_126 126 val_126 2008-04-08 11 126 val_126 126 val_126 2008-04-08 12 -126 val_126 126 val_126 2008-04-09 12 126 val_126 126 val_126 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 11 +126 val_126 126 val_126 2008-04-09 12 +128 val_128 128 val_128 2008-04-08 11 +128 val_128 128 val_128 2008-04-08 11 +128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 @@ -3718,355 +3162,355 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 128 val_128 128 val_128 2008-04-08 12 128 val_128 128 val_128 2008-04-08 12 128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 12 128 val_128 128 val_128 2008-04-09 12 128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 12 -129 val_129 129 val_129 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-08 11 129 val_129 129 val_129 2008-04-08 12 129 val_129 129 val_129 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 12 -129 val_129 129 val_129 2008-04-09 12 129 val_129 129 val_129 2008-04-08 12 129 val_129 129 val_129 2008-04-08 12 129 val_129 129 val_129 2008-04-09 11 129 val_129 129 val_129 2008-04-09 11 -129 val_129 129 val_129 2008-04-08 11 -129 val_129 129 val_129 2008-04-08 11 129 val_129 129 val_129 2008-04-09 11 129 val_129 129 val_129 2008-04-09 11 -129 val_129 129 val_129 2008-04-08 11 -129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +131 val_131 131 val_131 2008-04-08 11 131 val_131 131 val_131 2008-04-08 12 -131 val_131 131 val_131 2008-04-09 12 131 val_131 131 val_131 2008-04-09 11 -131 val_131 131 val_131 2008-04-08 11 +131 val_131 131 val_131 2008-04-09 12 133 val_133 133 val_133 2008-04-08 11 133 val_133 133 val_133 2008-04-08 12 133 val_133 133 val_133 2008-04-09 11 133 val_133 133 val_133 2008-04-09 12 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 12 +134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-09 11 -134 val_134 134 val_134 2008-04-08 11 -134 val_134 134 val_134 2008-04-08 11 134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-09 12 134 val_134 134 val_134 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 12 -134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-09 12 134 val_134 134 val_134 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 11 -134 val_134 134 val_134 2008-04-08 11 -136 val_136 136 val_136 2008-04-09 12 -136 val_136 136 val_136 2008-04-08 12 136 val_136 136 val_136 2008-04-08 11 +136 val_136 136 val_136 2008-04-08 12 136 val_136 136 val_136 2008-04-09 11 +136 val_136 136 val_136 2008-04-09 12 137 val_137 137 val_137 2008-04-08 11 137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-09 12 -137 val_137 137 val_137 2008-04-09 12 +137 val_137 137 val_137 2008-04-08 11 +137 val_137 137 val_137 2008-04-08 11 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 12 137 val_137 137 val_137 2008-04-09 12 -137 val_137 137 val_137 2008-04-08 12 -137 val_137 137 val_137 2008-04-08 12 -137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-08 12 -137 val_137 137 val_137 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 +137 val_137 137 val_137 2008-04-09 12 +137 val_137 137 val_137 2008-04-09 12 +138 val_138 138 val_138 2008-04-08 11 +138 val_138 138 val_138 2008-04-08 11 +138 val_138 138 val_138 2008-04-08 11 +138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 12 -143 val_143 143 val_143 2008-04-09 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 143 val_143 143 val_143 2008-04-08 11 -143 val_143 143 val_143 2008-04-09 11 143 val_143 143 val_143 2008-04-08 12 -145 val_145 145 val_145 2008-04-09 11 +143 val_143 143 val_143 2008-04-09 11 +143 val_143 143 val_143 2008-04-09 12 145 val_145 145 val_145 2008-04-08 11 145 val_145 145 val_145 2008-04-08 12 +145 val_145 145 val_145 2008-04-09 11 145 val_145 145 val_145 2008-04-09 12 146 val_146 146 val_146 2008-04-08 11 146 val_146 146 val_146 2008-04-08 11 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 -146 val_146 146 val_146 2008-04-08 12 146 val_146 146 val_146 2008-04-08 11 146 val_146 146 val_146 2008-04-08 11 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 146 val_146 146 val_146 2008-04-09 12 146 val_146 146 val_146 2008-04-09 12 146 val_146 146 val_146 2008-04-09 12 146 val_146 146 val_146 2008-04-09 12 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 -146 val_146 146 val_146 2008-04-08 12 -149 val_149 149 val_149 2008-04-08 12 -149 val_149 149 val_149 2008-04-08 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 11 -149 val_149 149 val_149 2008-04-09 11 -149 val_149 149 val_149 2008-04-09 11 -149 val_149 149 val_149 2008-04-09 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 12 149 val_149 149 val_149 2008-04-08 12 -150 val_150 150 val_150 2008-04-09 11 -150 val_150 150 val_150 2008-04-09 12 +149 val_149 149 val_149 2008-04-08 12 +149 val_149 149 val_149 2008-04-08 12 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 150 val_150 150 val_150 2008-04-08 11 150 val_150 150 val_150 2008-04-08 12 +150 val_150 150 val_150 2008-04-09 11 +150 val_150 150 val_150 2008-04-09 12 +152 val_152 152 val_152 2008-04-08 11 152 val_152 152 val_152 2008-04-08 11 152 val_152 152 val_152 2008-04-08 11 +152 val_152 152 val_152 2008-04-08 11 +152 val_152 152 val_152 2008-04-08 12 +152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-09 11 -152 val_152 152 val_152 2008-04-08 11 -152 val_152 152 val_152 2008-04-08 11 -152 val_152 152 val_152 2008-04-09 12 -152 val_152 152 val_152 2008-04-09 12 152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-09 12 152 val_152 152 val_152 2008-04-09 12 -152 val_152 152 val_152 2008-04-08 12 -152 val_152 152 val_152 2008-04-08 12 -153 val_153 153 val_153 2008-04-09 11 +152 val_152 152 val_152 2008-04-09 12 +152 val_152 152 val_152 2008-04-09 12 153 val_153 153 val_153 2008-04-08 11 153 val_153 153 val_153 2008-04-08 12 +153 val_153 153 val_153 2008-04-09 11 153 val_153 153 val_153 2008-04-09 12 -155 val_155 155 val_155 2008-04-08 12 155 val_155 155 val_155 2008-04-08 11 -155 val_155 155 val_155 2008-04-09 12 +155 val_155 155 val_155 2008-04-08 12 155 val_155 155 val_155 2008-04-09 11 +155 val_155 155 val_155 2008-04-09 12 +156 val_156 156 val_156 2008-04-08 11 156 val_156 156 val_156 2008-04-08 12 -156 val_156 156 val_156 2008-04-09 12 156 val_156 156 val_156 2008-04-09 11 -156 val_156 156 val_156 2008-04-08 11 -157 val_157 157 val_157 2008-04-09 11 -157 val_157 157 val_157 2008-04-08 12 +156 val_156 156 val_156 2008-04-09 12 157 val_157 157 val_157 2008-04-08 11 +157 val_157 157 val_157 2008-04-08 12 +157 val_157 157 val_157 2008-04-09 11 157 val_157 157 val_157 2008-04-09 12 158 val_158 158 val_158 2008-04-08 11 +158 val_158 158 val_158 2008-04-08 12 158 val_158 158 val_158 2008-04-09 11 158 val_158 158 val_158 2008-04-09 12 -158 val_158 158 val_158 2008-04-08 12 -160 val_160 160 val_160 2008-04-09 11 -160 val_160 160 val_160 2008-04-09 12 160 val_160 160 val_160 2008-04-08 11 160 val_160 160 val_160 2008-04-08 12 -162 val_162 162 val_162 2008-04-09 11 -162 val_162 162 val_162 2008-04-08 12 +160 val_160 160 val_160 2008-04-09 11 +160 val_160 160 val_160 2008-04-09 12 162 val_162 162 val_162 2008-04-08 11 +162 val_162 162 val_162 2008-04-08 12 +162 val_162 162 val_162 2008-04-09 11 162 val_162 162 val_162 2008-04-09 12 -163 val_163 163 val_163 2008-04-09 11 -163 val_163 163 val_163 2008-04-09 12 163 val_163 163 val_163 2008-04-08 11 163 val_163 163 val_163 2008-04-08 12 +163 val_163 163 val_163 2008-04-09 11 +163 val_163 163 val_163 2008-04-09 12 164 val_164 164 val_164 2008-04-08 11 164 val_164 164 val_164 2008-04-08 11 -164 val_164 164 val_164 2008-04-09 11 -164 val_164 164 val_164 2008-04-09 11 164 val_164 164 val_164 2008-04-08 11 164 val_164 164 val_164 2008-04-08 11 -164 val_164 164 val_164 2008-04-09 12 -164 val_164 164 val_164 2008-04-09 12 -164 val_164 164 val_164 2008-04-09 11 -164 val_164 164 val_164 2008-04-09 11 -164 val_164 164 val_164 2008-04-09 12 -164 val_164 164 val_164 2008-04-09 12 164 val_164 164 val_164 2008-04-08 12 164 val_164 164 val_164 2008-04-08 12 164 val_164 164 val_164 2008-04-08 12 164 val_164 164 val_164 2008-04-08 12 +164 val_164 164 val_164 2008-04-09 11 +164 val_164 164 val_164 2008-04-09 11 +164 val_164 164 val_164 2008-04-09 11 +164 val_164 164 val_164 2008-04-09 11 +164 val_164 164 val_164 2008-04-09 12 +164 val_164 164 val_164 2008-04-09 12 +164 val_164 164 val_164 2008-04-09 12 +164 val_164 164 val_164 2008-04-09 12 +165 val_165 165 val_165 2008-04-08 11 +165 val_165 165 val_165 2008-04-08 11 165 val_165 165 val_165 2008-04-08 11 165 val_165 165 val_165 2008-04-08 11 -165 val_165 165 val_165 2008-04-09 11 -165 val_165 165 val_165 2008-04-09 11 -165 val_165 165 val_165 2008-04-09 12 -165 val_165 165 val_165 2008-04-09 12 165 val_165 165 val_165 2008-04-08 12 165 val_165 165 val_165 2008-04-08 12 -165 val_165 165 val_165 2008-04-09 12 -165 val_165 165 val_165 2008-04-09 12 165 val_165 165 val_165 2008-04-08 12 165 val_165 165 val_165 2008-04-08 12 -165 val_165 165 val_165 2008-04-08 11 -165 val_165 165 val_165 2008-04-08 11 165 val_165 165 val_165 2008-04-09 11 165 val_165 165 val_165 2008-04-09 11 -166 val_166 166 val_166 2008-04-08 12 +165 val_165 165 val_165 2008-04-09 11 +165 val_165 165 val_165 2008-04-09 11 +165 val_165 165 val_165 2008-04-09 12 +165 val_165 165 val_165 2008-04-09 12 +165 val_165 165 val_165 2008-04-09 12 +165 val_165 165 val_165 2008-04-09 12 166 val_166 166 val_166 2008-04-08 11 -166 val_166 166 val_166 2008-04-09 12 +166 val_166 166 val_166 2008-04-08 12 166 val_166 166 val_166 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 +166 val_166 166 val_166 2008-04-09 12 +167 val_167 167 val_167 2008-04-08 11 +167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 +167 val_167 167 val_167 2008-04-08 11 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-08 12 168 val_168 168 val_168 2008-04-08 11 168 val_168 168 val_168 2008-04-08 12 168 val_168 168 val_168 2008-04-09 11 168 val_168 168 val_168 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 @@ -4079,6 +3523,21 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-09 11 @@ -4087,195 +3546,229 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 169 val_169 169 val_169 2008-04-09 12 169 val_169 169 val_169 2008-04-09 12 169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -170 val_170 170 val_170 2008-04-09 11 -170 val_170 170 val_170 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +17 val_17 17 val_17 2008-04-08 11 +17 val_17 17 val_17 2008-04-08 12 +17 val_17 17 val_17 2008-04-09 11 +17 val_17 17 val_17 2008-04-09 12 170 val_170 170 val_170 2008-04-08 11 170 val_170 170 val_170 2008-04-08 12 +170 val_170 170 val_170 2008-04-09 11 +170 val_170 170 val_170 2008-04-09 12 172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-08 11 -172 val_172 172 val_172 2008-04-09 11 -172 val_172 172 val_172 2008-04-09 11 -172 val_172 172 val_172 2008-04-08 12 -172 val_172 172 val_172 2008-04-08 12 -172 val_172 172 val_172 2008-04-09 12 -172 val_172 172 val_172 2008-04-09 12 172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-08 12 172 val_172 172 val_172 2008-04-08 12 -172 val_172 172 val_172 2008-04-09 12 -172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-08 12 +172 val_172 172 val_172 2008-04-08 12 172 val_172 172 val_172 2008-04-09 11 172 val_172 172 val_172 2008-04-09 11 -174 val_174 174 val_174 2008-04-08 12 -174 val_174 174 val_174 2008-04-08 12 -174 val_174 174 val_174 2008-04-09 12 -174 val_174 174 val_174 2008-04-09 12 -174 val_174 174 val_174 2008-04-09 12 -174 val_174 174 val_174 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 11 +172 val_172 172 val_172 2008-04-09 11 +172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 12 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 12 174 val_174 174 val_174 2008-04-08 12 +174 val_174 174 val_174 2008-04-08 12 +174 val_174 174 val_174 2008-04-08 12 174 val_174 174 val_174 2008-04-09 11 174 val_174 174 val_174 2008-04-09 11 174 val_174 174 val_174 2008-04-09 11 174 val_174 174 val_174 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 12 -175 val_175 175 val_175 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 12 -175 val_175 175 val_175 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 11 175 val_175 175 val_175 2008-04-08 12 175 val_175 175 val_175 2008-04-08 12 175 val_175 175 val_175 2008-04-08 12 175 val_175 175 val_175 2008-04-08 12 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 176 val_176 176 val_176 2008-04-08 11 176 val_176 176 val_176 2008-04-08 11 -176 val_176 176 val_176 2008-04-09 12 -176 val_176 176 val_176 2008-04-09 12 -176 val_176 176 val_176 2008-04-08 12 -176 val_176 176 val_176 2008-04-08 12 -176 val_176 176 val_176 2008-04-09 12 -176 val_176 176 val_176 2008-04-09 12 176 val_176 176 val_176 2008-04-08 11 176 val_176 176 val_176 2008-04-08 11 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 176 val_176 176 val_176 2008-04-09 11 176 val_176 176 val_176 2008-04-09 11 176 val_176 176 val_176 2008-04-09 11 176 val_176 176 val_176 2008-04-09 11 -176 val_176 176 val_176 2008-04-08 12 -176 val_176 176 val_176 2008-04-08 12 -177 val_177 177 val_177 2008-04-08 12 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 177 val_177 177 val_177 2008-04-08 11 +177 val_177 177 val_177 2008-04-08 12 177 val_177 177 val_177 2008-04-09 11 177 val_177 177 val_177 2008-04-09 12 +178 val_178 178 val_178 2008-04-08 11 178 val_178 178 val_178 2008-04-08 12 178 val_178 178 val_178 2008-04-09 11 -178 val_178 178 val_178 2008-04-08 11 178 val_178 178 val_178 2008-04-09 12 +179 val_179 179 val_179 2008-04-08 11 +179 val_179 179 val_179 2008-04-08 11 +179 val_179 179 val_179 2008-04-08 11 +179 val_179 179 val_179 2008-04-08 11 179 val_179 179 val_179 2008-04-08 12 179 val_179 179 val_179 2008-04-08 12 -179 val_179 179 val_179 2008-04-09 12 -179 val_179 179 val_179 2008-04-09 12 179 val_179 179 val_179 2008-04-08 12 179 val_179 179 val_179 2008-04-08 12 179 val_179 179 val_179 2008-04-09 11 179 val_179 179 val_179 2008-04-09 11 -179 val_179 179 val_179 2008-04-09 12 -179 val_179 179 val_179 2008-04-09 12 179 val_179 179 val_179 2008-04-09 11 179 val_179 179 val_179 2008-04-09 11 -179 val_179 179 val_179 2008-04-08 11 -179 val_179 179 val_179 2008-04-08 11 -179 val_179 179 val_179 2008-04-08 11 -179 val_179 179 val_179 2008-04-08 11 -180 val_180 180 val_180 2008-04-09 11 -180 val_180 180 val_180 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 180 val_180 180 val_180 2008-04-08 11 180 val_180 180 val_180 2008-04-08 12 -181 val_181 181 val_181 2008-04-09 11 -181 val_181 181 val_181 2008-04-09 12 +180 val_180 180 val_180 2008-04-09 11 +180 val_180 180 val_180 2008-04-09 12 181 val_181 181 val_181 2008-04-08 11 181 val_181 181 val_181 2008-04-08 12 -183 val_183 183 val_183 2008-04-08 12 +181 val_181 181 val_181 2008-04-09 11 +181 val_181 181 val_181 2008-04-09 12 183 val_183 183 val_183 2008-04-08 11 -183 val_183 183 val_183 2008-04-09 12 +183 val_183 183 val_183 2008-04-08 12 183 val_183 183 val_183 2008-04-09 11 +183 val_183 183 val_183 2008-04-09 12 186 val_186 186 val_186 2008-04-08 11 186 val_186 186 val_186 2008-04-08 12 186 val_186 186 val_186 2008-04-09 11 186 val_186 186 val_186 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 11 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 12 -189 val_189 189 val_189 2008-04-09 11 -189 val_189 189 val_189 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 189 val_189 189 val_189 2008-04-08 11 189 val_189 189 val_189 2008-04-08 12 -190 val_190 190 val_190 2008-04-09 12 +189 val_189 189 val_189 2008-04-09 11 +189 val_189 189 val_189 2008-04-09 12 +19 val_19 19 val_19 2008-04-08 11 +19 val_19 19 val_19 2008-04-08 12 +19 val_19 19 val_19 2008-04-09 11 +19 val_19 19 val_19 2008-04-09 12 +190 val_190 190 val_190 2008-04-08 11 190 val_190 190 val_190 2008-04-08 12 190 val_190 190 val_190 2008-04-09 11 -190 val_190 190 val_190 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-09 12 -191 val_191 191 val_191 2008-04-09 12 +190 val_190 190 val_190 2008-04-09 12 191 val_191 191 val_191 2008-04-08 11 191 val_191 191 val_191 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-08 12 -191 val_191 191 val_191 2008-04-08 12 191 val_191 191 val_191 2008-04-08 11 191 val_191 191 val_191 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 12 -191 val_191 191 val_191 2008-04-09 12 191 val_191 191 val_191 2008-04-08 12 191 val_191 191 val_191 2008-04-08 12 -192 val_192 192 val_192 2008-04-09 12 -192 val_192 192 val_192 2008-04-08 12 +191 val_191 191 val_191 2008-04-08 12 +191 val_191 191 val_191 2008-04-08 12 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-09 12 192 val_192 192 val_192 2008-04-08 11 +192 val_192 192 val_192 2008-04-08 12 192 val_192 192 val_192 2008-04-09 11 +192 val_192 192 val_192 2008-04-09 12 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 12 @@ -4291,82 +3784,70 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 11 -194 val_194 194 val_194 2008-04-08 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 194 val_194 194 val_194 2008-04-08 11 -194 val_194 194 val_194 2008-04-09 12 +194 val_194 194 val_194 2008-04-08 12 194 val_194 194 val_194 2008-04-09 11 +194 val_194 194 val_194 2008-04-09 12 195 val_195 195 val_195 2008-04-08 11 195 val_195 195 val_195 2008-04-08 11 -195 val_195 195 val_195 2008-04-09 11 -195 val_195 195 val_195 2008-04-09 11 -195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-08 12 -195 val_195 195 val_195 2008-04-08 12 -195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-09 12 195 val_195 195 val_195 2008-04-08 11 195 val_195 195 val_195 2008-04-08 11 195 val_195 195 val_195 2008-04-08 12 195 val_195 195 val_195 2008-04-08 12 +195 val_195 195 val_195 2008-04-08 12 +195 val_195 195 val_195 2008-04-08 12 +195 val_195 195 val_195 2008-04-09 11 195 val_195 195 val_195 2008-04-09 11 195 val_195 195 val_195 2008-04-09 11 +195 val_195 195 val_195 2008-04-09 11 +195 val_195 195 val_195 2008-04-09 12 +195 val_195 195 val_195 2008-04-09 12 +195 val_195 195 val_195 2008-04-09 12 +195 val_195 195 val_195 2008-04-09 12 +196 val_196 196 val_196 2008-04-08 11 196 val_196 196 val_196 2008-04-08 12 196 val_196 196 val_196 2008-04-09 11 -196 val_196 196 val_196 2008-04-08 11 196 val_196 196 val_196 2008-04-09 12 -197 val_197 197 val_197 2008-04-08 12 -197 val_197 197 val_197 2008-04-08 12 197 val_197 197 val_197 2008-04-08 11 197 val_197 197 val_197 2008-04-08 11 -197 val_197 197 val_197 2008-04-09 11 -197 val_197 197 val_197 2008-04-09 11 -197 val_197 197 val_197 2008-04-09 11 -197 val_197 197 val_197 2008-04-09 11 197 val_197 197 val_197 2008-04-08 11 197 val_197 197 val_197 2008-04-08 11 197 val_197 197 val_197 2008-04-08 12 197 val_197 197 val_197 2008-04-08 12 +197 val_197 197 val_197 2008-04-08 12 +197 val_197 197 val_197 2008-04-08 12 +197 val_197 197 val_197 2008-04-09 11 +197 val_197 197 val_197 2008-04-09 11 +197 val_197 197 val_197 2008-04-09 11 +197 val_197 197 val_197 2008-04-09 11 197 val_197 197 val_197 2008-04-09 12 197 val_197 197 val_197 2008-04-09 12 197 val_197 197 val_197 2008-04-09 12 197 val_197 197 val_197 2008-04-09 12 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 11 199 val_199 199 val_199 2008-04-08 11 199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 @@ -4376,301 +3857,315 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 12 -200 val_200 200 val_200 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +2 val_2 2 val_2 2008-04-08 11 +2 val_2 2 val_2 2008-04-08 12 +2 val_2 2 val_2 2008-04-09 11 +2 val_2 2 val_2 2008-04-09 12 +20 val_20 20 val_20 2008-04-08 11 +20 val_20 20 val_20 2008-04-08 12 +20 val_20 20 val_20 2008-04-09 11 +20 val_20 20 val_20 2008-04-09 12 +200 val_200 200 val_200 2008-04-08 11 +200 val_200 200 val_200 2008-04-08 11 200 val_200 200 val_200 2008-04-08 11 200 val_200 200 val_200 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 11 -200 val_200 200 val_200 2008-04-09 11 200 val_200 200 val_200 2008-04-08 12 200 val_200 200 val_200 2008-04-08 12 200 val_200 200 val_200 2008-04-08 12 200 val_200 200 val_200 2008-04-08 12 -200 val_200 200 val_200 2008-04-09 12 -200 val_200 200 val_200 2008-04-09 12 -200 val_200 200 val_200 2008-04-08 11 -200 val_200 200 val_200 2008-04-08 11 200 val_200 200 val_200 2008-04-09 11 200 val_200 200 val_200 2008-04-09 11 -201 val_201 201 val_201 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 201 val_201 201 val_201 2008-04-08 11 -201 val_201 201 val_201 2008-04-09 12 201 val_201 201 val_201 2008-04-08 12 -202 val_202 202 val_202 2008-04-09 12 +201 val_201 201 val_201 2008-04-09 11 +201 val_201 201 val_201 2008-04-09 12 202 val_202 202 val_202 2008-04-08 11 -202 val_202 202 val_202 2008-04-09 11 202 val_202 202 val_202 2008-04-08 12 -203 val_203 203 val_203 2008-04-09 11 -203 val_203 203 val_203 2008-04-09 11 +202 val_202 202 val_202 2008-04-09 11 +202 val_202 202 val_202 2008-04-09 12 +203 val_203 203 val_203 2008-04-08 11 203 val_203 203 val_203 2008-04-08 11 203 val_203 203 val_203 2008-04-08 11 +203 val_203 203 val_203 2008-04-08 11 +203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-08 12 203 val_203 203 val_203 2008-04-08 12 203 val_203 203 val_203 2008-04-08 12 203 val_203 203 val_203 2008-04-09 11 203 val_203 203 val_203 2008-04-09 11 -203 val_203 203 val_203 2008-04-08 11 -203 val_203 203 val_203 2008-04-08 11 +203 val_203 203 val_203 2008-04-09 11 +203 val_203 203 val_203 2008-04-09 11 203 val_203 203 val_203 2008-04-09 12 203 val_203 203 val_203 2008-04-09 12 203 val_203 203 val_203 2008-04-09 12 203 val_203 203 val_203 2008-04-09 12 -203 val_203 203 val_203 2008-04-08 12 -203 val_203 203 val_203 2008-04-08 12 +205 val_205 205 val_205 2008-04-08 11 +205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-08 12 205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 11 -205 val_205 205 val_205 2008-04-09 11 205 val_205 205 val_205 2008-04-08 12 205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 12 -205 val_205 205 val_205 2008-04-09 12 -205 val_205 205 val_205 2008-04-09 11 205 val_205 205 val_205 2008-04-09 11 -205 val_205 205 val_205 2008-04-08 11 -205 val_205 205 val_205 2008-04-08 11 -205 val_205 205 val_205 2008-04-09 12 -205 val_205 205 val_205 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 -207 val_207 207 val_207 2008-04-09 11 -207 val_207 207 val_207 2008-04-08 12 -207 val_207 207 val_207 2008-04-08 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 -207 val_207 207 val_207 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-09 12 207 val_207 207 val_207 2008-04-08 11 207 val_207 207 val_207 2008-04-08 11 -207 val_207 207 val_207 2008-04-08 12 -207 val_207 207 val_207 2008-04-08 12 207 val_207 207 val_207 2008-04-08 11 207 val_207 207 val_207 2008-04-08 11 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 12 +209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-09 11 209 val_209 209 val_209 2008-04-09 11 -209 val_209 209 val_209 2008-04-09 12 -209 val_209 209 val_209 2008-04-09 12 209 val_209 209 val_209 2008-04-09 11 209 val_209 209 val_209 2008-04-09 11 -209 val_209 209 val_209 2008-04-08 11 -209 val_209 209 val_209 2008-04-08 11 -209 val_209 209 val_209 2008-04-08 11 -209 val_209 209 val_209 2008-04-08 11 209 val_209 209 val_209 2008-04-09 12 209 val_209 209 val_209 2008-04-09 12 -209 val_209 209 val_209 2008-04-08 12 -209 val_209 209 val_209 2008-04-08 12 +209 val_209 209 val_209 2008-04-09 12 +209 val_209 209 val_209 2008-04-09 12 213 val_213 213 val_213 2008-04-08 11 213 val_213 213 val_213 2008-04-08 11 -213 val_213 213 val_213 2008-04-09 12 -213 val_213 213 val_213 2008-04-09 12 -213 val_213 213 val_213 2008-04-09 11 -213 val_213 213 val_213 2008-04-09 11 -213 val_213 213 val_213 2008-04-09 11 -213 val_213 213 val_213 2008-04-09 11 213 val_213 213 val_213 2008-04-08 11 213 val_213 213 val_213 2008-04-08 11 -213 val_213 213 val_213 2008-04-09 12 -213 val_213 213 val_213 2008-04-09 12 213 val_213 213 val_213 2008-04-08 12 213 val_213 213 val_213 2008-04-08 12 213 val_213 213 val_213 2008-04-08 12 213 val_213 213 val_213 2008-04-08 12 -214 val_214 214 val_214 2008-04-08 12 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-09 12 214 val_214 214 val_214 2008-04-08 11 +214 val_214 214 val_214 2008-04-08 12 214 val_214 214 val_214 2008-04-09 11 214 val_214 214 val_214 2008-04-09 12 -216 val_216 216 val_216 2008-04-09 11 -216 val_216 216 val_216 2008-04-09 11 -216 val_216 216 val_216 2008-04-08 12 -216 val_216 216 val_216 2008-04-08 12 216 val_216 216 val_216 2008-04-08 11 216 val_216 216 val_216 2008-04-08 11 -216 val_216 216 val_216 2008-04-09 12 -216 val_216 216 val_216 2008-04-09 12 +216 val_216 216 val_216 2008-04-08 11 +216 val_216 216 val_216 2008-04-08 11 +216 val_216 216 val_216 2008-04-08 12 216 val_216 216 val_216 2008-04-08 12 216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-09 11 +216 val_216 216 val_216 2008-04-09 11 216 val_216 216 val_216 2008-04-09 11 216 val_216 216 val_216 2008-04-09 11 216 val_216 216 val_216 2008-04-09 12 216 val_216 216 val_216 2008-04-09 12 -216 val_216 216 val_216 2008-04-08 11 -216 val_216 216 val_216 2008-04-08 11 -217 val_217 217 val_217 2008-04-08 12 -217 val_217 217 val_217 2008-04-08 12 +216 val_216 216 val_216 2008-04-09 12 +216 val_216 216 val_216 2008-04-09 12 217 val_217 217 val_217 2008-04-08 11 217 val_217 217 val_217 2008-04-08 11 -217 val_217 217 val_217 2008-04-09 11 -217 val_217 217 val_217 2008-04-09 11 -217 val_217 217 val_217 2008-04-08 12 -217 val_217 217 val_217 2008-04-08 12 217 val_217 217 val_217 2008-04-08 11 217 val_217 217 val_217 2008-04-08 11 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-09 11 +217 val_217 217 val_217 2008-04-09 11 217 val_217 217 val_217 2008-04-09 11 217 val_217 217 val_217 2008-04-09 11 217 val_217 217 val_217 2008-04-09 12 217 val_217 217 val_217 2008-04-09 12 217 val_217 217 val_217 2008-04-09 12 217 val_217 217 val_217 2008-04-09 12 -218 val_218 218 val_218 2008-04-08 12 218 val_218 218 val_218 2008-04-08 11 -218 val_218 218 val_218 2008-04-09 12 +218 val_218 218 val_218 2008-04-08 12 218 val_218 218 val_218 2008-04-09 11 +218 val_218 218 val_218 2008-04-09 12 219 val_219 219 val_219 2008-04-08 11 219 val_219 219 val_219 2008-04-08 11 -219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-09 11 -219 val_219 219 val_219 2008-04-09 11 -219 val_219 219 val_219 2008-04-08 12 -219 val_219 219 val_219 2008-04-08 12 219 val_219 219 val_219 2008-04-08 11 219 val_219 219 val_219 2008-04-08 11 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-09 11 +219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 12 219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-08 12 -219 val_219 219 val_219 2008-04-08 12 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-08 12 -221 val_221 221 val_221 2008-04-08 12 +219 val_219 219 val_219 2008-04-09 12 +219 val_219 219 val_219 2008-04-09 12 221 val_221 221 val_221 2008-04-08 11 221 val_221 221 val_221 2008-04-08 11 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-09 11 -221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-08 11 +221 val_221 221 val_221 2008-04-08 11 +221 val_221 221 val_221 2008-04-08 12 +221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-09 11 221 val_221 221 val_221 2008-04-09 11 -221 val_221 221 val_221 2008-04-08 11 -221 val_221 221 val_221 2008-04-08 11 +221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +222 val_222 222 val_222 2008-04-08 11 222 val_222 222 val_222 2008-04-08 12 -222 val_222 222 val_222 2008-04-09 12 222 val_222 222 val_222 2008-04-09 11 -222 val_222 222 val_222 2008-04-08 11 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-09 11 -223 val_223 223 val_223 2008-04-09 11 +222 val_222 222 val_222 2008-04-09 12 +223 val_223 223 val_223 2008-04-08 11 +223 val_223 223 val_223 2008-04-08 11 +223 val_223 223 val_223 2008-04-08 11 +223 val_223 223 val_223 2008-04-08 11 223 val_223 223 val_223 2008-04-08 12 223 val_223 223 val_223 2008-04-08 12 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-09 12 223 val_223 223 val_223 2008-04-08 12 223 val_223 223 val_223 2008-04-08 12 -223 val_223 223 val_223 2008-04-08 11 -223 val_223 223 val_223 2008-04-08 11 223 val_223 223 val_223 2008-04-09 11 223 val_223 223 val_223 2008-04-09 11 -223 val_223 223 val_223 2008-04-08 11 -223 val_223 223 val_223 2008-04-08 11 -224 val_224 224 val_224 2008-04-09 11 -224 val_224 224 val_224 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +224 val_224 224 val_224 2008-04-08 11 224 val_224 224 val_224 2008-04-08 11 224 val_224 224 val_224 2008-04-08 11 +224 val_224 224 val_224 2008-04-08 11 +224 val_224 224 val_224 2008-04-08 12 224 val_224 224 val_224 2008-04-08 12 224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 224 val_224 224 val_224 2008-04-09 12 224 val_224 224 val_224 2008-04-09 12 224 val_224 224 val_224 2008-04-09 12 224 val_224 224 val_224 2008-04-09 12 -224 val_224 224 val_224 2008-04-09 11 -224 val_224 224 val_224 2008-04-09 11 -224 val_224 224 val_224 2008-04-08 12 -224 val_224 224 val_224 2008-04-08 12 -224 val_224 224 val_224 2008-04-08 11 -224 val_224 224 val_224 2008-04-08 11 -226 val_226 226 val_226 2008-04-09 12 226 val_226 226 val_226 2008-04-08 11 -226 val_226 226 val_226 2008-04-09 11 226 val_226 226 val_226 2008-04-08 12 -228 val_228 228 val_228 2008-04-09 11 +226 val_226 226 val_226 2008-04-09 11 +226 val_226 226 val_226 2008-04-09 12 228 val_228 228 val_228 2008-04-08 11 -228 val_228 228 val_228 2008-04-09 12 228 val_228 228 val_228 2008-04-08 12 -229 val_229 229 val_229 2008-04-08 12 -229 val_229 229 val_229 2008-04-08 12 -229 val_229 229 val_229 2008-04-09 12 -229 val_229 229 val_229 2008-04-09 12 +228 val_228 228 val_228 2008-04-09 11 +228 val_228 228 val_228 2008-04-09 12 229 val_229 229 val_229 2008-04-08 11 229 val_229 229 val_229 2008-04-08 11 -229 val_229 229 val_229 2008-04-09 11 -229 val_229 229 val_229 2008-04-09 11 +229 val_229 229 val_229 2008-04-08 11 +229 val_229 229 val_229 2008-04-08 11 +229 val_229 229 val_229 2008-04-08 12 229 val_229 229 val_229 2008-04-08 12 229 val_229 229 val_229 2008-04-08 12 +229 val_229 229 val_229 2008-04-08 12 +229 val_229 229 val_229 2008-04-09 11 +229 val_229 229 val_229 2008-04-09 11 229 val_229 229 val_229 2008-04-09 11 229 val_229 229 val_229 2008-04-09 11 229 val_229 229 val_229 2008-04-09 12 229 val_229 229 val_229 2008-04-09 12 -229 val_229 229 val_229 2008-04-08 11 -229 val_229 229 val_229 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 +229 val_229 229 val_229 2008-04-09 12 +229 val_229 229 val_229 2008-04-09 12 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 230 val_230 230 val_230 2008-04-08 11 230 val_230 230 val_230 2008-04-08 11 230 val_230 230 val_230 2008-04-08 11 @@ -4681,41 +4176,46 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-09 11 @@ -4731,34 +4231,21 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 11 -233 val_233 233 val_233 2008-04-09 11 -233 val_233 233 val_233 2008-04-09 11 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 11 -233 val_233 233 val_233 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 233 val_233 233 val_233 2008-04-08 11 233 val_233 233 val_233 2008-04-08 11 233 val_233 233 val_233 2008-04-08 11 @@ -4767,118 +4254,138 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 233 val_233 233 val_233 2008-04-08 12 233 val_233 233 val_233 2008-04-08 12 233 val_233 233 val_233 2008-04-08 12 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 12 +233 val_233 233 val_233 2008-04-09 12 +233 val_233 233 val_233 2008-04-09 12 +233 val_233 233 val_233 2008-04-09 12 +235 val_235 235 val_235 2008-04-08 11 +235 val_235 235 val_235 2008-04-08 12 235 val_235 235 val_235 2008-04-09 11 235 val_235 235 val_235 2008-04-09 12 -235 val_235 235 val_235 2008-04-08 12 -235 val_235 235 val_235 2008-04-08 11 -237 val_237 237 val_237 2008-04-08 12 -237 val_237 237 val_237 2008-04-08 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 11 -237 val_237 237 val_237 2008-04-09 11 -237 val_237 237 val_237 2008-04-08 12 -237 val_237 237 val_237 2008-04-08 12 237 val_237 237 val_237 2008-04-08 11 237 val_237 237 val_237 2008-04-08 11 237 val_237 237 val_237 2008-04-08 11 237 val_237 237 val_237 2008-04-08 11 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 237 val_237 237 val_237 2008-04-09 11 237 val_237 237 val_237 2008-04-09 11 -238 val_238 238 val_238 2008-04-09 11 -238 val_238 238 val_238 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 238 val_238 238 val_238 2008-04-08 11 238 val_238 238 val_238 2008-04-08 11 -238 val_238 238 val_238 2008-04-09 12 -238 val_238 238 val_238 2008-04-09 12 238 val_238 238 val_238 2008-04-08 11 238 val_238 238 val_238 2008-04-08 11 238 val_238 238 val_238 2008-04-08 12 238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-09 11 +238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-09 12 238 val_238 238 val_238 2008-04-09 12 -238 val_238 238 val_238 2008-04-08 12 -238 val_238 238 val_238 2008-04-08 12 -239 val_239 239 val_239 2008-04-08 12 -239 val_239 239 val_239 2008-04-08 12 -239 val_239 239 val_239 2008-04-09 11 -239 val_239 239 val_239 2008-04-09 11 -239 val_239 239 val_239 2008-04-09 12 -239 val_239 239 val_239 2008-04-09 12 +238 val_238 238 val_238 2008-04-09 12 +238 val_238 238 val_238 2008-04-09 12 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 12 239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-09 11 +239 val_239 239 val_239 2008-04-09 11 239 val_239 239 val_239 2008-04-09 11 239 val_239 239 val_239 2008-04-09 11 239 val_239 239 val_239 2008-04-09 12 239 val_239 239 val_239 2008-04-09 12 -241 val_241 241 val_241 2008-04-08 12 -241 val_241 241 val_241 2008-04-09 12 +239 val_239 239 val_239 2008-04-09 12 +239 val_239 239 val_239 2008-04-09 12 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 241 val_241 241 val_241 2008-04-08 11 +241 val_241 241 val_241 2008-04-08 12 241 val_241 241 val_241 2008-04-09 11 +241 val_241 241 val_241 2008-04-09 12 +242 val_242 242 val_242 2008-04-08 11 +242 val_242 242 val_242 2008-04-08 11 242 val_242 242 val_242 2008-04-08 11 242 val_242 242 val_242 2008-04-08 11 -242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-09 12 -242 val_242 242 val_242 2008-04-09 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-09 11 242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-08 11 -242 val_242 242 val_242 2008-04-08 11 +242 val_242 242 val_242 2008-04-09 11 +242 val_242 242 val_242 2008-04-09 11 +242 val_242 242 val_242 2008-04-09 12 +242 val_242 242 val_242 2008-04-09 12 242 val_242 242 val_242 2008-04-09 12 242 val_242 242 val_242 2008-04-09 12 -244 val_244 244 val_244 2008-04-09 11 244 val_244 244 val_244 2008-04-08 11 244 val_244 244 val_244 2008-04-08 12 +244 val_244 244 val_244 2008-04-09 11 244 val_244 244 val_244 2008-04-09 12 247 val_247 247 val_247 2008-04-08 11 247 val_247 247 val_247 2008-04-08 12 247 val_247 247 val_247 2008-04-09 11 247 val_247 247 val_247 2008-04-09 12 248 val_248 248 val_248 2008-04-08 11 +248 val_248 248 val_248 2008-04-08 12 248 val_248 248 val_248 2008-04-09 11 248 val_248 248 val_248 2008-04-09 12 -248 val_248 248 val_248 2008-04-08 12 -249 val_249 249 val_249 2008-04-09 12 -249 val_249 249 val_249 2008-04-09 11 249 val_249 249 val_249 2008-04-08 11 249 val_249 249 val_249 2008-04-08 12 -252 val_252 252 val_252 2008-04-09 12 -252 val_252 252 val_252 2008-04-09 11 +249 val_249 249 val_249 2008-04-09 11 +249 val_249 249 val_249 2008-04-09 12 252 val_252 252 val_252 2008-04-08 11 252 val_252 252 val_252 2008-04-08 12 +252 val_252 252 val_252 2008-04-09 11 +252 val_252 252 val_252 2008-04-09 12 255 val_255 255 val_255 2008-04-08 11 255 val_255 255 val_255 2008-04-08 11 255 val_255 255 val_255 2008-04-08 11 255 val_255 255 val_255 2008-04-08 11 -255 val_255 255 val_255 2008-04-09 12 -255 val_255 255 val_255 2008-04-09 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-09 11 255 val_255 255 val_255 2008-04-09 11 -255 val_255 255 val_255 2008-04-09 12 -255 val_255 255 val_255 2008-04-09 12 255 val_255 255 val_255 2008-04-09 11 255 val_255 255 val_255 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 12 -256 val_256 256 val_256 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 256 val_256 256 val_256 2008-04-08 11 256 val_256 256 val_256 2008-04-08 11 256 val_256 256 val_256 2008-04-08 11 @@ -4889,140 +4396,160 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 256 val_256 256 val_256 2008-04-08 12 256 val_256 256 val_256 2008-04-09 11 256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 12 256 val_256 256 val_256 2008-04-09 12 256 val_256 256 val_256 2008-04-09 12 +256 val_256 256 val_256 2008-04-09 12 +257 val_257 257 val_257 2008-04-08 11 +257 val_257 257 val_257 2008-04-08 12 257 val_257 257 val_257 2008-04-09 11 257 val_257 257 val_257 2008-04-09 12 -257 val_257 257 val_257 2008-04-08 12 -257 val_257 257 val_257 2008-04-08 11 +258 val_258 258 val_258 2008-04-08 11 258 val_258 258 val_258 2008-04-08 12 258 val_258 258 val_258 2008-04-09 11 258 val_258 258 val_258 2008-04-09 12 -258 val_258 258 val_258 2008-04-08 11 -260 val_260 260 val_260 2008-04-09 12 -260 val_260 260 val_260 2008-04-09 11 -260 val_260 260 val_260 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 260 val_260 260 val_260 2008-04-08 11 +260 val_260 260 val_260 2008-04-08 12 +260 val_260 260 val_260 2008-04-09 11 +260 val_260 260 val_260 2008-04-09 12 262 val_262 262 val_262 2008-04-08 11 -262 val_262 262 val_262 2008-04-09 12 -262 val_262 262 val_262 2008-04-09 11 262 val_262 262 val_262 2008-04-08 12 -263 val_263 263 val_263 2008-04-09 11 +262 val_262 262 val_262 2008-04-09 11 +262 val_262 262 val_262 2008-04-09 12 263 val_263 263 val_263 2008-04-08 11 263 val_263 263 val_263 2008-04-08 12 +263 val_263 263 val_263 2008-04-09 11 263 val_263 263 val_263 2008-04-09 12 265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-08 11 +265 val_265 265 val_265 2008-04-08 11 +265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-08 12 265 val_265 265 val_265 2008-04-08 12 -265 val_265 265 val_265 2008-04-09 11 -265 val_265 265 val_265 2008-04-09 11 265 val_265 265 val_265 2008-04-08 12 265 val_265 265 val_265 2008-04-08 12 265 val_265 265 val_265 2008-04-09 11 265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-09 11 265 val_265 265 val_265 2008-04-09 12 265 val_265 265 val_265 2008-04-09 12 -265 val_265 265 val_265 2008-04-08 11 -265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-09 12 265 val_265 265 val_265 2008-04-09 12 +266 val_266 266 val_266 2008-04-08 11 266 val_266 266 val_266 2008-04-08 12 266 val_266 266 val_266 2008-04-09 11 -266 val_266 266 val_266 2008-04-08 11 266 val_266 266 val_266 2008-04-09 12 -272 val_272 272 val_272 2008-04-09 11 -272 val_272 272 val_272 2008-04-09 11 +27 val_27 27 val_27 2008-04-08 11 +27 val_27 27 val_27 2008-04-08 12 +27 val_27 27 val_27 2008-04-09 11 +27 val_27 27 val_27 2008-04-09 12 +272 val_272 272 val_272 2008-04-08 11 272 val_272 272 val_272 2008-04-08 11 272 val_272 272 val_272 2008-04-08 11 +272 val_272 272 val_272 2008-04-08 11 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-09 11 +272 val_272 272 val_272 2008-04-09 11 272 val_272 272 val_272 2008-04-09 11 272 val_272 272 val_272 2008-04-09 11 272 val_272 272 val_272 2008-04-09 12 272 val_272 272 val_272 2008-04-09 12 -272 val_272 272 val_272 2008-04-08 12 -272 val_272 272 val_272 2008-04-08 12 272 val_272 272 val_272 2008-04-09 12 272 val_272 272 val_272 2008-04-09 12 -272 val_272 272 val_272 2008-04-08 11 -272 val_272 272 val_272 2008-04-08 11 -272 val_272 272 val_272 2008-04-08 12 -272 val_272 272 val_272 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 12 273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 12 273 val_273 273 val_273 2008-04-09 12 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -274 val_274 274 val_274 2008-04-09 11 -274 val_274 274 val_274 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 274 val_274 274 val_274 2008-04-08 11 274 val_274 274 val_274 2008-04-08 12 -275 val_275 275 val_275 2008-04-09 12 +274 val_274 274 val_274 2008-04-09 11 +274 val_274 274 val_274 2008-04-09 12 275 val_275 275 val_275 2008-04-08 11 275 val_275 275 val_275 2008-04-08 12 275 val_275 275 val_275 2008-04-09 11 +275 val_275 275 val_275 2008-04-09 12 +277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 +277 val_277 277 val_277 2008-04-08 11 +277 val_277 277 val_277 2008-04-08 11 +277 val_277 277 val_277 2008-04-08 11 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 12 @@ -5039,10 +4566,6 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 @@ -5051,56 +4574,68 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 11 -278 val_278 278 val_278 2008-04-09 11 -278 val_278 278 val_278 2008-04-09 11 -278 val_278 278 val_278 2008-04-09 12 -278 val_278 278 val_278 2008-04-09 12 -278 val_278 278 val_278 2008-04-09 12 -278 val_278 278 val_278 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 12 278 val_278 278 val_278 2008-04-08 12 -278 val_278 278 val_278 2008-04-09 11 -278 val_278 278 val_278 2008-04-09 11 278 val_278 278 val_278 2008-04-08 12 278 val_278 278 val_278 2008-04-08 12 -280 val_280 280 val_280 2008-04-09 12 -280 val_280 280 val_280 2008-04-09 12 -280 val_280 280 val_280 2008-04-08 12 -280 val_280 280 val_280 2008-04-08 12 -280 val_280 280 val_280 2008-04-09 11 -280 val_280 280 val_280 2008-04-09 11 -280 val_280 280 val_280 2008-04-09 12 -280 val_280 280 val_280 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 12 +28 val_28 28 val_28 2008-04-08 11 +28 val_28 28 val_28 2008-04-08 12 +28 val_28 28 val_28 2008-04-09 11 +28 val_28 28 val_28 2008-04-09 12 280 val_280 280 val_280 2008-04-08 11 280 val_280 280 val_280 2008-04-08 11 -280 val_280 280 val_280 2008-04-08 12 -280 val_280 280 val_280 2008-04-08 12 280 val_280 280 val_280 2008-04-08 11 280 val_280 280 val_280 2008-04-08 11 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 280 val_280 280 val_280 2008-04-09 11 280 val_280 280 val_280 2008-04-09 11 -281 val_281 281 val_281 2008-04-09 12 -281 val_281 281 val_281 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 11 +280 val_280 280 val_280 2008-04-09 11 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 281 val_281 281 val_281 2008-04-08 11 281 val_281 281 val_281 2008-04-08 11 -281 val_281 281 val_281 2008-04-08 12 -281 val_281 281 val_281 2008-04-08 12 -281 val_281 281 val_281 2008-04-09 11 -281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-08 11 281 val_281 281 val_281 2008-04-08 11 281 val_281 281 val_281 2008-04-08 12 281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-09 11 +281 val_281 281 val_281 2008-04-09 11 +281 val_281 281 val_281 2008-04-09 12 +281 val_281 281 val_281 2008-04-09 12 281 val_281 281 val_281 2008-04-09 12 281 val_281 281 val_281 2008-04-09 12 282 val_282 282 val_282 2008-04-08 11 @@ -5109,180 +4644,166 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 282 val_282 282 val_282 2008-04-08 11 282 val_282 282 val_282 2008-04-08 12 282 val_282 282 val_282 2008-04-08 12 -282 val_282 282 val_282 2008-04-09 12 -282 val_282 282 val_282 2008-04-09 12 +282 val_282 282 val_282 2008-04-08 12 +282 val_282 282 val_282 2008-04-08 12 282 val_282 282 val_282 2008-04-09 11 282 val_282 282 val_282 2008-04-09 11 282 val_282 282 val_282 2008-04-09 11 282 val_282 282 val_282 2008-04-09 11 -282 val_282 282 val_282 2008-04-08 12 -282 val_282 282 val_282 2008-04-08 12 282 val_282 282 val_282 2008-04-09 12 282 val_282 282 val_282 2008-04-09 12 -283 val_283 283 val_283 2008-04-09 12 +282 val_282 282 val_282 2008-04-09 12 +282 val_282 282 val_282 2008-04-09 12 283 val_283 283 val_283 2008-04-08 11 283 val_283 283 val_283 2008-04-08 12 283 val_283 283 val_283 2008-04-09 11 +283 val_283 283 val_283 2008-04-09 12 284 val_284 284 val_284 2008-04-08 11 -284 val_284 284 val_284 2008-04-09 12 284 val_284 284 val_284 2008-04-08 12 284 val_284 284 val_284 2008-04-09 11 -285 val_285 285 val_285 2008-04-09 12 -285 val_285 285 val_285 2008-04-09 11 +284 val_284 284 val_284 2008-04-09 12 285 val_285 285 val_285 2008-04-08 11 285 val_285 285 val_285 2008-04-08 12 +285 val_285 285 val_285 2008-04-09 11 +285 val_285 285 val_285 2008-04-09 12 286 val_286 286 val_286 2008-04-08 11 286 val_286 286 val_286 2008-04-08 12 -286 val_286 286 val_286 2008-04-09 12 286 val_286 286 val_286 2008-04-09 11 +286 val_286 286 val_286 2008-04-09 12 287 val_287 287 val_287 2008-04-08 11 -287 val_287 287 val_287 2008-04-09 12 -287 val_287 287 val_287 2008-04-09 11 287 val_287 287 val_287 2008-04-08 12 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 12 -288 val_288 288 val_288 2008-04-09 12 -288 val_288 288 val_288 2008-04-08 12 -288 val_288 288 val_288 2008-04-08 12 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 12 -288 val_288 288 val_288 2008-04-09 12 +287 val_287 287 val_287 2008-04-09 11 +287 val_287 287 val_287 2008-04-09 12 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 12 288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 12 +288 val_288 288 val_288 2008-04-09 12 +288 val_288 288 val_288 2008-04-09 12 +288 val_288 288 val_288 2008-04-09 12 +289 val_289 289 val_289 2008-04-08 11 289 val_289 289 val_289 2008-04-08 12 289 val_289 289 val_289 2008-04-09 11 -289 val_289 289 val_289 2008-04-08 11 289 val_289 289 val_289 2008-04-09 12 -291 val_291 291 val_291 2008-04-09 12 +291 val_291 291 val_291 2008-04-08 11 291 val_291 291 val_291 2008-04-08 12 291 val_291 291 val_291 2008-04-09 11 -291 val_291 291 val_291 2008-04-08 11 -292 val_292 292 val_292 2008-04-09 11 +291 val_291 291 val_291 2008-04-09 12 292 val_292 292 val_292 2008-04-08 11 -292 val_292 292 val_292 2008-04-09 12 292 val_292 292 val_292 2008-04-08 12 -296 val_296 296 val_296 2008-04-08 12 +292 val_292 292 val_292 2008-04-09 11 +292 val_292 292 val_292 2008-04-09 12 296 val_296 296 val_296 2008-04-08 11 -296 val_296 296 val_296 2008-04-09 12 +296 val_296 296 val_296 2008-04-08 12 296 val_296 296 val_296 2008-04-09 11 +296 val_296 296 val_296 2008-04-09 12 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 11 298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-09 11 +298 val_298 298 val_298 2008-04-09 11 +298 val_298 298 val_298 2008-04-09 11 +298 val_298 298 val_298 2008-04-09 11 +298 val_298 298 val_298 2008-04-09 11 +298 val_298 298 val_298 2008-04-09 11 298 val_298 298 val_298 2008-04-09 11 298 val_298 298 val_298 2008-04-09 11 298 val_298 298 val_298 2008-04-09 11 298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-08 12 -302 val_302 302 val_302 2008-04-08 12 -302 val_302 302 val_302 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +30 val_30 30 val_30 2008-04-08 11 +30 val_30 30 val_30 2008-04-08 12 +30 val_30 30 val_30 2008-04-09 11 +30 val_30 30 val_30 2008-04-09 12 302 val_302 302 val_302 2008-04-08 11 +302 val_302 302 val_302 2008-04-08 12 302 val_302 302 val_302 2008-04-09 11 +302 val_302 302 val_302 2008-04-09 12 305 val_305 305 val_305 2008-04-08 11 -305 val_305 305 val_305 2008-04-09 12 -305 val_305 305 val_305 2008-04-09 11 305 val_305 305 val_305 2008-04-08 12 -306 val_306 306 val_306 2008-04-09 11 +305 val_305 305 val_305 2008-04-09 11 +305 val_305 305 val_305 2008-04-09 12 306 val_306 306 val_306 2008-04-08 11 -306 val_306 306 val_306 2008-04-09 12 306 val_306 306 val_306 2008-04-08 12 +306 val_306 306 val_306 2008-04-09 11 +306 val_306 306 val_306 2008-04-09 12 307 val_307 307 val_307 2008-04-08 11 307 val_307 307 val_307 2008-04-08 11 -307 val_307 307 val_307 2008-04-08 12 -307 val_307 307 val_307 2008-04-08 12 -307 val_307 307 val_307 2008-04-09 12 -307 val_307 307 val_307 2008-04-09 12 -307 val_307 307 val_307 2008-04-09 12 -307 val_307 307 val_307 2008-04-09 12 307 val_307 307 val_307 2008-04-08 11 307 val_307 307 val_307 2008-04-08 11 307 val_307 307 val_307 2008-04-08 12 307 val_307 307 val_307 2008-04-08 12 +307 val_307 307 val_307 2008-04-08 12 +307 val_307 307 val_307 2008-04-08 12 307 val_307 307 val_307 2008-04-09 11 307 val_307 307 val_307 2008-04-09 11 307 val_307 307 val_307 2008-04-09 11 307 val_307 307 val_307 2008-04-09 11 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 308 val_308 308 val_308 2008-04-08 11 -308 val_308 308 val_308 2008-04-09 12 308 val_308 308 val_308 2008-04-08 12 308 val_308 308 val_308 2008-04-09 11 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-09 11 -309 val_309 309 val_309 2008-04-09 11 +308 val_308 308 val_308 2008-04-09 12 309 val_309 309 val_309 2008-04-08 11 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-08 12 -309 val_309 309 val_309 2008-04-08 12 309 val_309 309 val_309 2008-04-08 11 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 11 -309 val_309 309 val_309 2008-04-09 11 309 val_309 309 val_309 2008-04-08 12 309 val_309 309 val_309 2008-04-08 12 -310 val_310 310 val_310 2008-04-09 11 -310 val_310 310 val_310 2008-04-09 12 +309 val_309 309 val_309 2008-04-08 12 +309 val_309 309 val_309 2008-04-08 12 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 310 val_310 310 val_310 2008-04-08 11 310 val_310 310 val_310 2008-04-08 12 +310 val_310 310 val_310 2008-04-09 11 +310 val_310 310 val_310 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 11 @@ -5295,86 +4816,89 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 315 val_315 315 val_315 2008-04-08 11 -315 val_315 315 val_315 2008-04-09 12 -315 val_315 315 val_315 2008-04-09 11 315 val_315 315 val_315 2008-04-08 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-08 12 +315 val_315 315 val_315 2008-04-09 11 +315 val_315 315 val_315 2008-04-09 12 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 316 val_316 316 val_316 2008-04-08 11 316 val_316 316 val_316 2008-04-08 11 316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 317 val_317 317 val_317 2008-04-08 11 317 val_317 317 val_317 2008-04-08 11 -317 val_317 317 val_317 2008-04-09 11 -317 val_317 317 val_317 2008-04-09 11 -317 val_317 317 val_317 2008-04-09 12 -317 val_317 317 val_317 2008-04-09 12 -317 val_317 317 val_317 2008-04-09 12 -317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-08 11 +317 val_317 317 val_317 2008-04-08 11 317 val_317 317 val_317 2008-04-08 12 317 val_317 317 val_317 2008-04-08 12 317 val_317 317 val_317 2008-04-08 12 317 val_317 317 val_317 2008-04-08 12 -317 val_317 317 val_317 2008-04-08 11 -317 val_317 317 val_317 2008-04-08 11 317 val_317 317 val_317 2008-04-09 11 317 val_317 317 val_317 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 11 +317 val_317 317 val_317 2008-04-09 11 +317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 12 +318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 @@ -5384,11 +4908,24 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-08 11 -321 val_321 321 val_321 2008-04-09 11 -321 val_321 321 val_321 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 321 val_321 321 val_321 2008-04-08 11 321 val_321 321 val_321 2008-04-08 11 321 val_321 321 val_321 2008-04-08 11 @@ -5399,46 +4936,54 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 321 val_321 321 val_321 2008-04-08 12 321 val_321 321 val_321 2008-04-09 11 321 val_321 321 val_321 2008-04-09 11 +321 val_321 321 val_321 2008-04-09 11 +321 val_321 321 val_321 2008-04-09 11 321 val_321 321 val_321 2008-04-09 12 321 val_321 321 val_321 2008-04-09 12 321 val_321 321 val_321 2008-04-09 12 321 val_321 321 val_321 2008-04-09 12 -322 val_322 322 val_322 2008-04-08 12 -322 val_322 322 val_322 2008-04-08 12 322 val_322 322 val_322 2008-04-08 11 322 val_322 322 val_322 2008-04-08 11 +322 val_322 322 val_322 2008-04-08 11 +322 val_322 322 val_322 2008-04-08 11 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-09 11 +322 val_322 322 val_322 2008-04-09 11 322 val_322 322 val_322 2008-04-09 11 322 val_322 322 val_322 2008-04-09 11 322 val_322 322 val_322 2008-04-09 12 322 val_322 322 val_322 2008-04-09 12 322 val_322 322 val_322 2008-04-09 12 322 val_322 322 val_322 2008-04-09 12 -322 val_322 322 val_322 2008-04-09 11 -322 val_322 322 val_322 2008-04-09 11 -322 val_322 322 val_322 2008-04-08 12 -322 val_322 322 val_322 2008-04-08 12 -322 val_322 322 val_322 2008-04-08 11 -322 val_322 322 val_322 2008-04-08 11 323 val_323 323 val_323 2008-04-08 11 323 val_323 323 val_323 2008-04-08 12 323 val_323 323 val_323 2008-04-09 11 323 val_323 323 val_323 2008-04-09 12 -325 val_325 325 val_325 2008-04-09 11 -325 val_325 325 val_325 2008-04-09 11 +325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-08 11 325 val_325 325 val_325 2008-04-08 12 325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-09 12 -325 val_325 325 val_325 2008-04-09 12 +325 val_325 325 val_325 2008-04-08 12 +325 val_325 325 val_325 2008-04-08 12 +325 val_325 325 val_325 2008-04-09 11 +325 val_325 325 val_325 2008-04-09 11 325 val_325 325 val_325 2008-04-09 11 325 val_325 325 val_325 2008-04-09 11 325 val_325 325 val_325 2008-04-09 12 325 val_325 325 val_325 2008-04-09 12 -325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-08 11 -325 val_325 325 val_325 2008-04-08 11 -325 val_325 325 val_325 2008-04-08 11 -325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-09 12 +325 val_325 325 val_325 2008-04-09 12 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-08 11 @@ -5448,6 +4993,10 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 @@ -5456,137 +5005,135 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 -331 val_331 331 val_331 2008-04-09 12 -331 val_331 331 val_331 2008-04-09 12 +33 val_33 33 val_33 2008-04-08 11 +33 val_33 33 val_33 2008-04-08 12 +33 val_33 33 val_33 2008-04-09 11 +33 val_33 33 val_33 2008-04-09 12 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 12 331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-08 12 331 val_331 331 val_331 2008-04-09 11 331 val_331 331 val_331 2008-04-09 11 331 val_331 331 val_331 2008-04-09 11 331 val_331 331 val_331 2008-04-09 11 -331 val_331 331 val_331 2008-04-08 12 -331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-09 12 +331 val_331 331 val_331 2008-04-09 12 331 val_331 331 val_331 2008-04-09 12 331 val_331 331 val_331 2008-04-09 12 332 val_332 332 val_332 2008-04-08 11 -332 val_332 332 val_332 2008-04-09 11 332 val_332 332 val_332 2008-04-08 12 +332 val_332 332 val_332 2008-04-09 11 332 val_332 332 val_332 2008-04-09 12 -333 val_333 333 val_333 2008-04-09 11 -333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-08 11 333 val_333 333 val_333 2008-04-08 12 333 val_333 333 val_333 2008-04-08 12 -333 val_333 333 val_333 2008-04-09 12 -333 val_333 333 val_333 2008-04-09 12 333 val_333 333 val_333 2008-04-08 12 333 val_333 333 val_333 2008-04-08 12 -333 val_333 333 val_333 2008-04-08 11 -333 val_333 333 val_333 2008-04-08 11 333 val_333 333 val_333 2008-04-09 11 333 val_333 333 val_333 2008-04-09 11 -333 val_333 333 val_333 2008-04-08 11 -333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-09 12 333 val_333 333 val_333 2008-04-09 12 333 val_333 333 val_333 2008-04-09 12 +333 val_333 333 val_333 2008-04-09 12 +335 val_335 335 val_335 2008-04-08 11 335 val_335 335 val_335 2008-04-08 12 -335 val_335 335 val_335 2008-04-09 12 335 val_335 335 val_335 2008-04-09 11 -335 val_335 335 val_335 2008-04-08 11 +335 val_335 335 val_335 2008-04-09 12 +336 val_336 336 val_336 2008-04-08 11 +336 val_336 336 val_336 2008-04-08 12 336 val_336 336 val_336 2008-04-09 11 336 val_336 336 val_336 2008-04-09 12 -336 val_336 336 val_336 2008-04-08 12 -336 val_336 336 val_336 2008-04-08 11 338 val_338 338 val_338 2008-04-08 11 -338 val_338 338 val_338 2008-04-09 11 338 val_338 338 val_338 2008-04-08 12 +338 val_338 338 val_338 2008-04-09 11 338 val_338 338 val_338 2008-04-09 12 -339 val_339 339 val_339 2008-04-09 11 339 val_339 339 val_339 2008-04-08 11 -339 val_339 339 val_339 2008-04-09 12 339 val_339 339 val_339 2008-04-08 12 -341 val_341 341 val_341 2008-04-09 12 -341 val_341 341 val_341 2008-04-09 11 +339 val_339 339 val_339 2008-04-09 11 +339 val_339 339 val_339 2008-04-09 12 +34 val_34 34 val_34 2008-04-08 11 +34 val_34 34 val_34 2008-04-08 12 +34 val_34 34 val_34 2008-04-09 11 +34 val_34 34 val_34 2008-04-09 12 341 val_341 341 val_341 2008-04-08 11 341 val_341 341 val_341 2008-04-08 12 +341 val_341 341 val_341 2008-04-09 11 +341 val_341 341 val_341 2008-04-09 12 +342 val_342 342 val_342 2008-04-08 11 +342 val_342 342 val_342 2008-04-08 11 342 val_342 342 val_342 2008-04-08 11 342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 -342 val_342 342 val_342 2008-04-09 12 342 val_342 342 val_342 2008-04-08 12 342 val_342 342 val_342 2008-04-08 12 -342 val_342 342 val_342 2008-04-09 11 -342 val_342 342 val_342 2008-04-09 11 342 val_342 342 val_342 2008-04-08 12 342 val_342 342 val_342 2008-04-08 12 -342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 -342 val_342 342 val_342 2008-04-09 12 342 val_342 342 val_342 2008-04-09 11 342 val_342 342 val_342 2008-04-09 11 -344 val_344 344 val_344 2008-04-08 12 -344 val_344 344 val_344 2008-04-08 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 11 -344 val_344 344 val_344 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 +344 val_344 344 val_344 2008-04-08 11 344 val_344 344 val_344 2008-04-08 11 344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-08 12 +344 val_344 344 val_344 2008-04-08 12 344 val_344 344 val_344 2008-04-08 12 344 val_344 344 val_344 2008-04-08 12 344 val_344 344 val_344 2008-04-09 11 344 val_344 344 val_344 2008-04-09 11 -344 val_344 344 val_344 2008-04-08 11 -344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-09 11 +344 val_344 344 val_344 2008-04-09 11 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 345 val_345 345 val_345 2008-04-08 11 -345 val_345 345 val_345 2008-04-09 11 345 val_345 345 val_345 2008-04-08 12 +345 val_345 345 val_345 2008-04-09 11 345 val_345 345 val_345 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-08 11 @@ -5597,11 +5144,36 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 348 val_348 348 val_348 2008-04-08 12 348 val_348 348 val_348 2008-04-08 12 348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 @@ -5612,291 +5184,303 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 11 -351 val_351 351 val_351 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 351 val_351 351 val_351 2008-04-08 11 351 val_351 351 val_351 2008-04-08 12 +351 val_351 351 val_351 2008-04-09 11 351 val_351 351 val_351 2008-04-09 12 +353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 11 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-09 11 353 val_353 353 val_353 2008-04-09 11 -353 val_353 353 val_353 2008-04-09 12 -353 val_353 353 val_353 2008-04-09 12 -353 val_353 353 val_353 2008-04-08 11 -353 val_353 353 val_353 2008-04-08 11 353 val_353 353 val_353 2008-04-09 11 353 val_353 353 val_353 2008-04-09 11 353 val_353 353 val_353 2008-04-09 12 353 val_353 353 val_353 2008-04-09 12 -353 val_353 353 val_353 2008-04-08 11 -353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-09 12 +353 val_353 353 val_353 2008-04-09 12 +356 val_356 356 val_356 2008-04-08 11 356 val_356 356 val_356 2008-04-08 12 356 val_356 356 val_356 2008-04-09 11 -356 val_356 356 val_356 2008-04-08 11 356 val_356 356 val_356 2008-04-09 12 360 val_360 360 val_360 2008-04-08 11 360 val_360 360 val_360 2008-04-08 12 -360 val_360 360 val_360 2008-04-09 12 360 val_360 360 val_360 2008-04-09 11 -362 val_362 362 val_362 2008-04-09 12 +360 val_360 360 val_360 2008-04-09 12 362 val_362 362 val_362 2008-04-08 11 -362 val_362 362 val_362 2008-04-09 11 362 val_362 362 val_362 2008-04-08 12 -364 val_364 364 val_364 2008-04-09 12 +362 val_362 362 val_362 2008-04-09 11 +362 val_362 362 val_362 2008-04-09 12 364 val_364 364 val_364 2008-04-08 11 -364 val_364 364 val_364 2008-04-09 11 364 val_364 364 val_364 2008-04-08 12 +364 val_364 364 val_364 2008-04-09 11 +364 val_364 364 val_364 2008-04-09 12 365 val_365 365 val_365 2008-04-08 11 -365 val_365 365 val_365 2008-04-09 12 365 val_365 365 val_365 2008-04-08 12 365 val_365 365 val_365 2008-04-09 11 +365 val_365 365 val_365 2008-04-09 12 +366 val_366 366 val_366 2008-04-08 11 366 val_366 366 val_366 2008-04-08 12 -366 val_366 366 val_366 2008-04-09 12 366 val_366 366 val_366 2008-04-09 11 -366 val_366 366 val_366 2008-04-08 11 -367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 11 -367 val_367 367 val_367 2008-04-09 11 +366 val_366 366 val_366 2008-04-09 12 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 12 367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-08 12 +367 val_367 367 val_367 2008-04-08 12 367 val_367 367 val_367 2008-04-09 11 367 val_367 367 val_367 2008-04-09 11 -368 val_368 368 val_368 2008-04-08 12 +367 val_367 367 val_367 2008-04-09 11 +367 val_367 367 val_367 2008-04-09 11 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 368 val_368 368 val_368 2008-04-08 11 -368 val_368 368 val_368 2008-04-09 12 +368 val_368 368 val_368 2008-04-08 12 368 val_368 368 val_368 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-08 12 +368 val_368 368 val_368 2008-04-09 12 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 12 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 12 369 val_369 369 val_369 2008-04-08 12 369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 373 val_373 373 val_373 2008-04-08 11 +373 val_373 373 val_373 2008-04-08 12 373 val_373 373 val_373 2008-04-09 11 373 val_373 373 val_373 2008-04-09 12 -373 val_373 373 val_373 2008-04-08 12 -374 val_374 374 val_374 2008-04-09 12 -374 val_374 374 val_374 2008-04-08 12 374 val_374 374 val_374 2008-04-08 11 +374 val_374 374 val_374 2008-04-08 12 374 val_374 374 val_374 2008-04-09 11 -375 val_375 375 val_375 2008-04-09 12 -375 val_375 375 val_375 2008-04-09 11 -375 val_375 375 val_375 2008-04-08 12 +374 val_374 374 val_374 2008-04-09 12 375 val_375 375 val_375 2008-04-08 11 -377 val_377 377 val_377 2008-04-09 11 -377 val_377 377 val_377 2008-04-09 12 +375 val_375 375 val_375 2008-04-08 12 +375 val_375 375 val_375 2008-04-09 11 +375 val_375 375 val_375 2008-04-09 12 377 val_377 377 val_377 2008-04-08 11 377 val_377 377 val_377 2008-04-08 12 +377 val_377 377 val_377 2008-04-09 11 +377 val_377 377 val_377 2008-04-09 12 +378 val_378 378 val_378 2008-04-08 11 +378 val_378 378 val_378 2008-04-08 12 378 val_378 378 val_378 2008-04-09 11 378 val_378 378 val_378 2008-04-09 12 -378 val_378 378 val_378 2008-04-08 12 -378 val_378 378 val_378 2008-04-08 11 -379 val_379 379 val_379 2008-04-08 12 379 val_379 379 val_379 2008-04-08 11 +379 val_379 379 val_379 2008-04-08 12 379 val_379 379 val_379 2008-04-09 11 379 val_379 379 val_379 2008-04-09 12 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 12 +382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-09 11 382 val_382 382 val_382 2008-04-09 11 382 val_382 382 val_382 2008-04-09 11 382 val_382 382 val_382 2008-04-09 11 -382 val_382 382 val_382 2008-04-08 12 -382 val_382 382 val_382 2008-04-08 12 -382 val_382 382 val_382 2008-04-08 11 -382 val_382 382 val_382 2008-04-08 11 382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-09 12 -382 val_382 382 val_382 2008-04-08 11 -382 val_382 382 val_382 2008-04-08 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +386 val_386 386 val_386 2008-04-08 11 386 val_386 386 val_386 2008-04-08 12 -386 val_386 386 val_386 2008-04-09 12 386 val_386 386 val_386 2008-04-09 11 -386 val_386 386 val_386 2008-04-08 11 -389 val_389 389 val_389 2008-04-08 12 -389 val_389 389 val_389 2008-04-09 12 +386 val_386 386 val_386 2008-04-09 12 389 val_389 389 val_389 2008-04-08 11 +389 val_389 389 val_389 2008-04-08 12 389 val_389 389 val_389 2008-04-09 11 -392 val_392 392 val_392 2008-04-08 12 +389 val_389 389 val_389 2008-04-09 12 392 val_392 392 val_392 2008-04-08 11 -392 val_392 392 val_392 2008-04-09 12 +392 val_392 392 val_392 2008-04-08 12 392 val_392 392 val_392 2008-04-09 11 -393 val_393 393 val_393 2008-04-09 12 -393 val_393 393 val_393 2008-04-08 12 +392 val_392 392 val_392 2008-04-09 12 393 val_393 393 val_393 2008-04-08 11 +393 val_393 393 val_393 2008-04-08 12 393 val_393 393 val_393 2008-04-09 11 -394 val_394 394 val_394 2008-04-08 12 -394 val_394 394 val_394 2008-04-09 12 +393 val_393 393 val_393 2008-04-09 12 394 val_394 394 val_394 2008-04-08 11 +394 val_394 394 val_394 2008-04-08 12 394 val_394 394 val_394 2008-04-09 11 -395 val_395 395 val_395 2008-04-09 12 -395 val_395 395 val_395 2008-04-09 12 +394 val_394 394 val_394 2008-04-09 12 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 395 val_395 395 val_395 2008-04-09 11 395 val_395 395 val_395 2008-04-09 11 395 val_395 395 val_395 2008-04-09 11 395 val_395 395 val_395 2008-04-09 11 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 11 -395 val_395 395 val_395 2008-04-08 11 395 val_395 395 val_395 2008-04-09 12 395 val_395 395 val_395 2008-04-09 12 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 11 -395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-09 12 +395 val_395 395 val_395 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 @@ -5906,23 +5490,30 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 396 val_396 396 val_396 2008-04-08 12 396 val_396 396 val_396 2008-04-08 12 396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 396 val_396 396 val_396 2008-04-08 12 396 val_396 396 val_396 2008-04-08 12 396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 11 -397 val_397 397 val_397 2008-04-09 11 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 11 -397 val_397 397 val_397 2008-04-09 11 397 val_397 397 val_397 2008-04-08 11 397 val_397 397 val_397 2008-04-08 11 397 val_397 397 val_397 2008-04-08 11 @@ -5931,26 +5522,78 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 397 val_397 397 val_397 2008-04-08 12 397 val_397 397 val_397 2008-04-08 12 397 val_397 397 val_397 2008-04-08 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 11 -399 val_399 399 val_399 2008-04-09 11 -399 val_399 399 val_399 2008-04-08 12 -399 val_399 399 val_399 2008-04-08 12 -399 val_399 399 val_399 2008-04-09 11 -399 val_399 399 val_399 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 12 +397 val_397 397 val_397 2008-04-09 12 +397 val_397 397 val_397 2008-04-09 12 +397 val_397 397 val_397 2008-04-09 12 399 val_399 399 val_399 2008-04-08 11 399 val_399 399 val_399 2008-04-08 11 -399 val_399 399 val_399 2008-04-08 12 -399 val_399 399 val_399 2008-04-08 12 399 val_399 399 val_399 2008-04-08 11 399 val_399 399 val_399 2008-04-08 11 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +4 val_4 4 val_4 2008-04-08 11 +4 val_4 4 val_4 2008-04-08 12 +4 val_4 4 val_4 2008-04-09 11 +4 val_4 4 val_4 2008-04-09 12 +400 val_400 400 val_400 2008-04-08 11 400 val_400 400 val_400 2008-04-08 12 400 val_400 400 val_400 2008-04-09 11 400 val_400 400 val_400 2008-04-09 12 -400 val_400 400 val_400 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 @@ -5976,21 +5619,6 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 @@ -6006,11 +5634,6 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 @@ -6021,61 +5644,35 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 -402 val_402 402 val_402 2008-04-09 11 402 val_402 402 val_402 2008-04-08 11 -402 val_402 402 val_402 2008-04-09 12 402 val_402 402 val_402 2008-04-08 12 +402 val_402 402 val_402 2008-04-09 11 +402 val_402 402 val_402 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 403 val_403 403 val_403 2008-04-08 12 403 val_403 403 val_403 2008-04-08 12 403 val_403 403 val_403 2008-04-08 12 @@ -6085,32 +5682,66 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 403 val_403 403 val_403 2008-04-09 11 403 val_403 403 val_403 2008-04-09 11 403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 11 -404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 12 404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-09 11 +404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-09 12 404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-08 12 -404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-09 12 +404 val_404 404 val_404 2008-04-09 12 +406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 @@ -6119,10 +5750,6 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 @@ -6147,208 +5774,211 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 407 val_407 407 val_407 2008-04-08 11 407 val_407 407 val_407 2008-04-08 12 -407 val_407 407 val_407 2008-04-09 12 407 val_407 407 val_407 2008-04-09 11 +407 val_407 407 val_407 2008-04-09 12 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +41 val_41 41 val_41 2008-04-08 11 +41 val_41 41 val_41 2008-04-08 12 +41 val_41 41 val_41 2008-04-09 11 +41 val_41 41 val_41 2008-04-09 12 411 val_411 411 val_411 2008-04-08 11 411 val_411 411 val_411 2008-04-08 12 -411 val_411 411 val_411 2008-04-09 12 411 val_411 411 val_411 2008-04-09 11 -413 val_413 413 val_413 2008-04-09 11 -413 val_413 413 val_413 2008-04-09 11 +411 val_411 411 val_411 2008-04-09 12 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 12 413 val_413 413 val_413 2008-04-08 12 413 val_413 413 val_413 2008-04-08 12 +413 val_413 413 val_413 2008-04-08 12 +413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-09 11 413 val_413 413 val_413 2008-04-09 11 413 val_413 413 val_413 2008-04-09 11 413 val_413 413 val_413 2008-04-09 12 413 val_413 413 val_413 2008-04-09 12 -413 val_413 413 val_413 2008-04-08 12 -413 val_413 413 val_413 2008-04-08 12 -413 val_413 413 val_413 2008-04-08 11 -413 val_413 413 val_413 2008-04-08 11 413 val_413 413 val_413 2008-04-09 12 413 val_413 413 val_413 2008-04-09 12 -413 val_413 413 val_413 2008-04-08 11 -413 val_413 413 val_413 2008-04-08 11 414 val_414 414 val_414 2008-04-08 11 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-08 12 -414 val_414 414 val_414 2008-04-08 12 -414 val_414 414 val_414 2008-04-09 11 -414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-08 11 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-09 12 -414 val_414 414 val_414 2008-04-09 12 414 val_414 414 val_414 2008-04-08 12 414 val_414 414 val_414 2008-04-08 12 +414 val_414 414 val_414 2008-04-08 12 +414 val_414 414 val_414 2008-04-08 12 +414 val_414 414 val_414 2008-04-09 11 +414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-09 12 414 val_414 414 val_414 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-08 12 +414 val_414 414 val_414 2008-04-09 12 +414 val_414 414 val_414 2008-04-09 12 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +418 val_418 418 val_418 2008-04-08 11 418 val_418 418 val_418 2008-04-08 12 -418 val_418 418 val_418 2008-04-09 12 418 val_418 418 val_418 2008-04-09 11 -418 val_418 418 val_418 2008-04-08 11 -419 val_419 419 val_419 2008-04-08 12 +418 val_418 418 val_418 2008-04-09 12 419 val_419 419 val_419 2008-04-08 11 -419 val_419 419 val_419 2008-04-09 12 +419 val_419 419 val_419 2008-04-08 12 419 val_419 419 val_419 2008-04-09 11 +419 val_419 419 val_419 2008-04-09 12 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 421 val_421 421 val_421 2008-04-08 11 421 val_421 421 val_421 2008-04-08 12 421 val_421 421 val_421 2008-04-09 11 421 val_421 421 val_421 2008-04-09 12 424 val_424 424 val_424 2008-04-08 11 424 val_424 424 val_424 2008-04-08 11 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-08 11 +424 val_424 424 val_424 2008-04-08 11 424 val_424 424 val_424 2008-04-08 12 424 val_424 424 val_424 2008-04-08 12 424 val_424 424 val_424 2008-04-08 12 424 val_424 424 val_424 2008-04-08 12 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 424 val_424 424 val_424 2008-04-09 12 424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-08 11 -424 val_424 424 val_424 2008-04-08 11 -427 val_427 427 val_427 2008-04-09 12 +424 val_424 424 val_424 2008-04-09 12 +424 val_424 424 val_424 2008-04-09 12 +427 val_427 427 val_427 2008-04-08 11 427 val_427 427 val_427 2008-04-08 12 427 val_427 427 val_427 2008-04-09 11 -427 val_427 427 val_427 2008-04-08 11 -429 val_429 429 val_429 2008-04-09 11 -429 val_429 429 val_429 2008-04-09 11 +427 val_427 427 val_427 2008-04-09 12 429 val_429 429 val_429 2008-04-08 11 429 val_429 429 val_429 2008-04-08 11 -429 val_429 429 val_429 2008-04-08 12 -429 val_429 429 val_429 2008-04-08 12 -429 val_429 429 val_429 2008-04-09 12 -429 val_429 429 val_429 2008-04-09 12 429 val_429 429 val_429 2008-04-08 11 429 val_429 429 val_429 2008-04-08 11 -429 val_429 429 val_429 2008-04-09 12 -429 val_429 429 val_429 2008-04-09 12 429 val_429 429 val_429 2008-04-08 12 429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-09 11 429 val_429 429 val_429 2008-04-09 11 429 val_429 429 val_429 2008-04-09 11 +429 val_429 429 val_429 2008-04-09 11 +429 val_429 429 val_429 2008-04-09 12 +429 val_429 429 val_429 2008-04-09 12 +429 val_429 429 val_429 2008-04-09 12 +429 val_429 429 val_429 2008-04-09 12 +43 val_43 43 val_43 2008-04-08 11 +43 val_43 43 val_43 2008-04-08 12 +43 val_43 43 val_43 2008-04-09 11 +43 val_43 43 val_43 2008-04-09 12 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 430 val_430 430 val_430 2008-04-08 11 430 val_430 430 val_430 2008-04-08 11 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-09 11 @@ -6358,27 +5988,21 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 @@ -6388,138 +6012,148 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -432 val_432 432 val_432 2008-04-09 12 432 val_432 432 val_432 2008-04-08 11 -432 val_432 432 val_432 2008-04-09 11 432 val_432 432 val_432 2008-04-08 12 -435 val_435 435 val_435 2008-04-09 11 +432 val_432 432 val_432 2008-04-09 11 +432 val_432 432 val_432 2008-04-09 12 435 val_435 435 val_435 2008-04-08 11 435 val_435 435 val_435 2008-04-08 12 +435 val_435 435 val_435 2008-04-09 11 435 val_435 435 val_435 2008-04-09 12 -436 val_436 436 val_436 2008-04-08 12 436 val_436 436 val_436 2008-04-08 11 +436 val_436 436 val_436 2008-04-08 12 436 val_436 436 val_436 2008-04-09 11 436 val_436 436 val_436 2008-04-09 12 -437 val_437 437 val_437 2008-04-09 12 437 val_437 437 val_437 2008-04-08 11 -437 val_437 437 val_437 2008-04-09 11 437 val_437 437 val_437 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-09 11 +437 val_437 437 val_437 2008-04-09 11 +437 val_437 437 val_437 2008-04-09 12 +438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 +438 val_438 438 val_438 2008-04-08 11 +438 val_438 438 val_438 2008-04-08 11 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-09 12 438 val_438 438 val_438 2008-04-09 12 438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 11 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-08 11 439 val_439 439 val_439 2008-04-08 12 439 val_439 439 val_439 2008-04-08 12 439 val_439 439 val_439 2008-04-08 12 439 val_439 439 val_439 2008-04-08 12 -439 val_439 439 val_439 2008-04-08 11 -439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 439 val_439 439 val_439 2008-04-09 12 439 val_439 439 val_439 2008-04-09 12 -439 val_439 439 val_439 2008-04-08 11 -439 val_439 439 val_439 2008-04-08 11 439 val_439 439 val_439 2008-04-09 12 439 val_439 439 val_439 2008-04-09 12 -443 val_443 443 val_443 2008-04-09 11 -443 val_443 443 val_443 2008-04-09 12 +44 val_44 44 val_44 2008-04-08 11 +44 val_44 44 val_44 2008-04-08 12 +44 val_44 44 val_44 2008-04-09 11 +44 val_44 44 val_44 2008-04-09 12 443 val_443 443 val_443 2008-04-08 11 443 val_443 443 val_443 2008-04-08 12 -444 val_444 444 val_444 2008-04-09 11 +443 val_443 443 val_443 2008-04-09 11 +443 val_443 443 val_443 2008-04-09 12 444 val_444 444 val_444 2008-04-08 11 444 val_444 444 val_444 2008-04-08 12 +444 val_444 444 val_444 2008-04-09 11 444 val_444 444 val_444 2008-04-09 12 -446 val_446 446 val_446 2008-04-08 12 -446 val_446 446 val_446 2008-04-09 12 446 val_446 446 val_446 2008-04-08 11 +446 val_446 446 val_446 2008-04-08 12 446 val_446 446 val_446 2008-04-09 11 -448 val_448 448 val_448 2008-04-09 11 -448 val_448 448 val_448 2008-04-09 12 +446 val_446 446 val_446 2008-04-09 12 448 val_448 448 val_448 2008-04-08 11 448 val_448 448 val_448 2008-04-08 12 -449 val_449 449 val_449 2008-04-08 12 +448 val_448 448 val_448 2008-04-09 11 +448 val_448 448 val_448 2008-04-09 12 449 val_449 449 val_449 2008-04-08 11 -449 val_449 449 val_449 2008-04-09 12 +449 val_449 449 val_449 2008-04-08 12 449 val_449 449 val_449 2008-04-09 11 +449 val_449 449 val_449 2008-04-09 12 452 val_452 452 val_452 2008-04-08 11 +452 val_452 452 val_452 2008-04-08 12 452 val_452 452 val_452 2008-04-09 11 452 val_452 452 val_452 2008-04-09 12 -452 val_452 452 val_452 2008-04-08 12 +453 val_453 453 val_453 2008-04-08 11 453 val_453 453 val_453 2008-04-08 12 453 val_453 453 val_453 2008-04-09 11 453 val_453 453 val_453 2008-04-09 12 -453 val_453 453 val_453 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-08 12 454 val_454 454 val_454 2008-04-08 12 454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-09 11 @@ -6529,106 +6163,109 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 455 val_455 455 val_455 2008-04-08 11 -455 val_455 455 val_455 2008-04-09 11 455 val_455 455 val_455 2008-04-08 12 +455 val_455 455 val_455 2008-04-09 11 455 val_455 455 val_455 2008-04-09 12 -457 val_457 457 val_457 2008-04-08 12 -457 val_457 457 val_457 2008-04-09 12 457 val_457 457 val_457 2008-04-08 11 +457 val_457 457 val_457 2008-04-08 12 457 val_457 457 val_457 2008-04-09 11 +457 val_457 457 val_457 2008-04-09 12 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 12 +458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-09 11 458 val_458 458 val_458 2008-04-09 11 -458 val_458 458 val_458 2008-04-08 11 -458 val_458 458 val_458 2008-04-08 11 458 val_458 458 val_458 2008-04-09 11 458 val_458 458 val_458 2008-04-09 11 -458 val_458 458 val_458 2008-04-08 11 -458 val_458 458 val_458 2008-04-08 11 458 val_458 458 val_458 2008-04-09 12 458 val_458 458 val_458 2008-04-09 12 458 val_458 458 val_458 2008-04-09 12 458 val_458 458 val_458 2008-04-09 12 -458 val_458 458 val_458 2008-04-08 12 -458 val_458 458 val_458 2008-04-08 12 +459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-08 12 +459 val_459 459 val_459 2008-04-08 12 459 val_459 459 val_459 2008-04-08 12 459 val_459 459 val_459 2008-04-08 12 -459 val_459 459 val_459 2008-04-09 12 -459 val_459 459 val_459 2008-04-09 12 459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-09 11 -459 val_459 459 val_459 2008-04-08 11 -459 val_459 459 val_459 2008-04-08 11 -459 val_459 459 val_459 2008-04-08 12 -459 val_459 459 val_459 2008-04-08 12 459 val_459 459 val_459 2008-04-09 12 459 val_459 459 val_459 2008-04-09 12 -459 val_459 459 val_459 2008-04-08 11 -459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-09 12 +459 val_459 459 val_459 2008-04-09 12 460 val_460 460 val_460 2008-04-08 11 -460 val_460 460 val_460 2008-04-09 11 460 val_460 460 val_460 2008-04-08 12 +460 val_460 460 val_460 2008-04-09 11 460 val_460 460 val_460 2008-04-09 12 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 12 -462 val_462 462 val_462 2008-04-09 12 462 val_462 462 val_462 2008-04-08 11 462 val_462 462 val_462 2008-04-08 11 462 val_462 462 val_462 2008-04-08 11 462 val_462 462 val_462 2008-04-08 11 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 12 -462 val_462 462 val_462 2008-04-09 12 462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-08 12 -463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-08 12 -463 val_463 463 val_463 2008-04-08 12 +462 val_462 462 val_462 2008-04-09 11 +462 val_462 462 val_462 2008-04-09 11 +462 val_462 462 val_462 2008-04-09 11 +462 val_462 462 val_462 2008-04-09 11 +462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-09 12 +463 val_463 463 val_463 2008-04-08 11 463 val_463 463 val_463 2008-04-08 11 463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-08 12 +463 val_463 463 val_463 2008-04-08 12 463 val_463 463 val_463 2008-04-08 12 463 val_463 463 val_463 2008-04-08 12 463 val_463 463 val_463 2008-04-09 11 463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-08 11 -463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-09 11 +463 val_463 463 val_463 2008-04-09 11 +463 val_463 463 val_463 2008-04-09 12 +463 val_463 463 val_463 2008-04-09 12 +463 val_463 463 val_463 2008-04-09 12 +463 val_463 463 val_463 2008-04-09 12 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 @@ -6638,35 +6275,33 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-09 12 -467 val_467 467 val_467 2008-04-09 12 -467 val_467 467 val_467 2008-04-09 11 -467 val_467 467 val_467 2008-04-08 12 467 val_467 467 val_467 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 +467 val_467 467 val_467 2008-04-08 12 +467 val_467 467 val_467 2008-04-09 11 +467 val_467 467 val_467 2008-04-09 12 +468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 12 @@ -6683,30 +6318,10 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 @@ -6715,21 +6330,47 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 @@ -6740,21 +6381,6 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 @@ -6765,11 +6391,38 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 @@ -6780,80 +6433,59 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +47 val_47 47 val_47 2008-04-08 11 +47 val_47 47 val_47 2008-04-08 12 +47 val_47 47 val_47 2008-04-09 11 +47 val_47 47 val_47 2008-04-09 12 470 val_470 470 val_470 2008-04-08 11 -470 val_470 470 val_470 2008-04-09 12 470 val_470 470 val_470 2008-04-08 12 470 val_470 470 val_470 2008-04-09 11 -472 val_472 472 val_472 2008-04-09 11 +470 val_470 470 val_470 2008-04-09 12 +472 val_472 472 val_472 2008-04-08 11 472 val_472 472 val_472 2008-04-08 12 +472 val_472 472 val_472 2008-04-09 11 472 val_472 472 val_472 2008-04-09 12 -472 val_472 472 val_472 2008-04-08 11 -475 val_475 475 val_475 2008-04-09 11 475 val_475 475 val_475 2008-04-08 11 475 val_475 475 val_475 2008-04-08 12 +475 val_475 475 val_475 2008-04-09 11 475 val_475 475 val_475 2008-04-09 12 -477 val_477 477 val_477 2008-04-09 12 477 val_477 477 val_477 2008-04-08 11 477 val_477 477 val_477 2008-04-08 12 477 val_477 477 val_477 2008-04-09 11 +477 val_477 477 val_477 2008-04-09 12 +478 val_478 478 val_478 2008-04-08 11 478 val_478 478 val_478 2008-04-08 11 478 val_478 478 val_478 2008-04-08 11 +478 val_478 478 val_478 2008-04-08 11 +478 val_478 478 val_478 2008-04-08 12 478 val_478 478 val_478 2008-04-08 12 478 val_478 478 val_478 2008-04-08 12 +478 val_478 478 val_478 2008-04-08 12 +478 val_478 478 val_478 2008-04-09 11 +478 val_478 478 val_478 2008-04-09 11 478 val_478 478 val_478 2008-04-09 11 478 val_478 478 val_478 2008-04-09 11 478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-09 12 -478 val_478 478 val_478 2008-04-08 11 -478 val_478 478 val_478 2008-04-08 11 -478 val_478 478 val_478 2008-04-09 11 -478 val_478 478 val_478 2008-04-09 11 -478 val_478 478 val_478 2008-04-08 12 -478 val_478 478 val_478 2008-04-08 12 -479 val_479 479 val_479 2008-04-08 12 479 val_479 479 val_479 2008-04-08 11 +479 val_479 479 val_479 2008-04-08 12 479 val_479 479 val_479 2008-04-09 11 479 val_479 479 val_479 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 @@ -6863,15 +6495,12 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 @@ -6884,41 +6513,39 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 480 val_480 480 val_480 2008-04-09 11 480 val_480 480 val_480 2008-04-09 11 480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-08 12 -481 val_481 481 val_481 2008-04-08 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 481 val_481 481 val_481 2008-04-08 11 +481 val_481 481 val_481 2008-04-08 12 481 val_481 481 val_481 2008-04-09 11 481 val_481 481 val_481 2008-04-09 12 -482 val_482 482 val_482 2008-04-09 11 -482 val_482 482 val_482 2008-04-09 12 482 val_482 482 val_482 2008-04-08 11 482 val_482 482 val_482 2008-04-08 12 -483 val_483 483 val_483 2008-04-09 11 +482 val_482 482 val_482 2008-04-09 11 +482 val_482 482 val_482 2008-04-09 12 483 val_483 483 val_483 2008-04-08 11 -483 val_483 483 val_483 2008-04-09 12 483 val_483 483 val_483 2008-04-08 12 -484 val_484 484 val_484 2008-04-08 12 +483 val_483 483 val_483 2008-04-09 11 +483 val_483 483 val_483 2008-04-09 12 484 val_484 484 val_484 2008-04-08 11 +484 val_484 484 val_484 2008-04-08 12 484 val_484 484 val_484 2008-04-09 11 484 val_484 484 val_484 2008-04-09 12 485 val_485 485 val_485 2008-04-08 11 +485 val_485 485 val_485 2008-04-08 12 485 val_485 485 val_485 2008-04-09 11 485 val_485 485 val_485 2008-04-09 12 -485 val_485 485 val_485 2008-04-08 12 -487 val_487 487 val_487 2008-04-09 12 +487 val_487 487 val_487 2008-04-08 11 487 val_487 487 val_487 2008-04-08 12 487 val_487 487 val_487 2008-04-09 11 -487 val_487 487 val_487 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 +487 val_487 487 val_487 2008-04-09 12 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 @@ -6931,42 +6558,22 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 @@ -6975,86 +6582,458 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 489 val_489 489 val_489 2008-04-09 11 489 val_489 489 val_489 2008-04-09 11 489 val_489 489 val_489 2008-04-09 11 -490 val_490 490 val_490 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 490 val_490 490 val_490 2008-04-08 11 +490 val_490 490 val_490 2008-04-08 12 +490 val_490 490 val_490 2008-04-09 11 490 val_490 490 val_490 2008-04-09 12 -490 val_490 490 val_490 2008-04-08 12 491 val_491 491 val_491 2008-04-08 11 491 val_491 491 val_491 2008-04-08 12 491 val_491 491 val_491 2008-04-09 11 491 val_491 491 val_491 2008-04-09 12 492 val_492 492 val_492 2008-04-08 11 492 val_492 492 val_492 2008-04-08 11 +492 val_492 492 val_492 2008-04-08 11 +492 val_492 492 val_492 2008-04-08 11 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-09 11 +492 val_492 492 val_492 2008-04-09 11 492 val_492 492 val_492 2008-04-09 11 492 val_492 492 val_492 2008-04-09 11 492 val_492 492 val_492 2008-04-09 12 492 val_492 492 val_492 2008-04-09 12 492 val_492 492 val_492 2008-04-09 12 492 val_492 492 val_492 2008-04-09 12 -492 val_492 492 val_492 2008-04-08 12 -492 val_492 492 val_492 2008-04-08 12 -492 val_492 492 val_492 2008-04-09 11 -492 val_492 492 val_492 2008-04-09 11 -492 val_492 492 val_492 2008-04-08 11 -492 val_492 492 val_492 2008-04-08 11 -492 val_492 492 val_492 2008-04-08 12 -492 val_492 492 val_492 2008-04-08 12 +493 val_493 493 val_493 2008-04-08 11 493 val_493 493 val_493 2008-04-08 12 -493 val_493 493 val_493 2008-04-09 12 493 val_493 493 val_493 2008-04-09 11 -493 val_493 493 val_493 2008-04-08 11 -494 val_494 494 val_494 2008-04-09 11 +493 val_493 493 val_493 2008-04-09 12 494 val_494 494 val_494 2008-04-08 11 494 val_494 494 val_494 2008-04-08 12 +494 val_494 494 val_494 2008-04-09 11 494 val_494 494 val_494 2008-04-09 12 495 val_495 495 val_495 2008-04-08 11 -495 val_495 495 val_495 2008-04-09 12 495 val_495 495 val_495 2008-04-08 12 495 val_495 495 val_495 2008-04-09 11 +495 val_495 495 val_495 2008-04-09 12 +496 val_496 496 val_496 2008-04-08 11 +496 val_496 496 val_496 2008-04-08 12 496 val_496 496 val_496 2008-04-09 11 496 val_496 496 val_496 2008-04-09 12 -496 val_496 496 val_496 2008-04-08 12 -496 val_496 496 val_496 2008-04-08 11 -497 val_497 497 val_497 2008-04-09 12 -497 val_497 497 val_497 2008-04-08 12 497 val_497 497 val_497 2008-04-08 11 +497 val_497 497 val_497 2008-04-08 12 497 val_497 497 val_497 2008-04-09 11 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 +497 val_497 497 val_497 2008-04-09 12 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +53 val_53 53 val_53 2008-04-08 11 +53 val_53 53 val_53 2008-04-08 12 +53 val_53 53 val_53 2008-04-09 11 +53 val_53 53 val_53 2008-04-09 12 +54 val_54 54 val_54 2008-04-08 11 +54 val_54 54 val_54 2008-04-08 12 +54 val_54 54 val_54 2008-04-09 11 +54 val_54 54 val_54 2008-04-09 12 +57 val_57 57 val_57 2008-04-08 11 +57 val_57 57 val_57 2008-04-08 12 +57 val_57 57 val_57 2008-04-09 11 +57 val_57 57 val_57 2008-04-09 12 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +64 val_64 64 val_64 2008-04-08 11 +64 val_64 64 val_64 2008-04-08 12 +64 val_64 64 val_64 2008-04-09 11 +64 val_64 64 val_64 2008-04-09 12 +65 val_65 65 val_65 2008-04-08 11 +65 val_65 65 val_65 2008-04-08 12 +65 val_65 65 val_65 2008-04-09 11 +65 val_65 65 val_65 2008-04-09 12 +66 val_66 66 val_66 2008-04-08 11 +66 val_66 66 val_66 2008-04-08 12 +66 val_66 66 val_66 2008-04-09 11 +66 val_66 66 val_66 2008-04-09 12 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +69 val_69 69 val_69 2008-04-08 11 +69 val_69 69 val_69 2008-04-08 12 +69 val_69 69 val_69 2008-04-09 11 +69 val_69 69 val_69 2008-04-09 12 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +74 val_74 74 val_74 2008-04-08 11 +74 val_74 74 val_74 2008-04-08 12 +74 val_74 74 val_74 2008-04-09 11 +74 val_74 74 val_74 2008-04-09 12 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +77 val_77 77 val_77 2008-04-08 11 +77 val_77 77 val_77 2008-04-08 12 +77 val_77 77 val_77 2008-04-09 11 +77 val_77 77 val_77 2008-04-09 12 +78 val_78 78 val_78 2008-04-08 11 +78 val_78 78 val_78 2008-04-08 12 +78 val_78 78 val_78 2008-04-09 11 +78 val_78 78 val_78 2008-04-09 12 +8 val_8 8 val_8 2008-04-08 11 +8 val_8 8 val_8 2008-04-08 12 +8 val_8 8 val_8 2008-04-09 11 +8 val_8 8 val_8 2008-04-09 12 +80 val_80 80 val_80 2008-04-08 11 +80 val_80 80 val_80 2008-04-08 12 +80 val_80 80 val_80 2008-04-09 11 +80 val_80 80 val_80 2008-04-09 12 +82 val_82 82 val_82 2008-04-08 11 +82 val_82 82 val_82 2008-04-08 12 +82 val_82 82 val_82 2008-04-09 11 +82 val_82 82 val_82 2008-04-09 12 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +85 val_85 85 val_85 2008-04-08 11 +85 val_85 85 val_85 2008-04-08 12 +85 val_85 85 val_85 2008-04-09 11 +85 val_85 85 val_85 2008-04-09 12 +86 val_86 86 val_86 2008-04-08 11 +86 val_86 86 val_86 2008-04-08 12 +86 val_86 86 val_86 2008-04-09 11 +86 val_86 86 val_86 2008-04-09 12 +87 val_87 87 val_87 2008-04-08 11 +87 val_87 87 val_87 2008-04-08 12 +87 val_87 87 val_87 2008-04-09 11 +87 val_87 87 val_87 2008-04-09 12 +9 val_9 9 val_9 2008-04-08 11 +9 val_9 9 val_9 2008-04-08 12 +9 val_9 9 val_9 2008-04-09 11 +9 val_9 9 val_9 2008-04-09 12 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +92 val_92 92 val_92 2008-04-08 11 +92 val_92 92 val_92 2008-04-08 12 +92 val_92 92 val_92 2008-04-09 11 +92 val_92 92 val_92 2008-04-09 12 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +96 val_96 96 val_96 2008-04-08 11 +96 val_96 96 val_96 2008-04-08 12 +96 val_96 96 val_96 2008-04-09 11 +96 val_96 96 val_96 2008-04-09 12 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 PREHOOK: query: explain select * from default.masking_test_subq_n3 where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq_n3 @@ -7066,151 +7045,150 @@ POSTHOOK: Input: default@masking_test_subq_n3 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 + Stage-1 is a root stage Stage-0 depends on stages: Stage-1 STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) Stage: Stage-0 Fetch Operator @@ -7225,90 +7203,10 @@ PREHOOK: Input: default@src #### A masked pattern was here #### POSTHOOK: query: select * from default.masking_test_subq_n3 where key > 0 POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n3 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -2 val_2 -4 val_4 -5 val_5 -5 val_5 -5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -80 val_80 -82 val_82 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -87 val_87 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -98 val_98 -98 val_98 +POSTHOOK: Input: default@masking_test_subq_n3 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 100 val_100 100 val_100 103 val_103 @@ -7316,6 +7214,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -7326,6 +7225,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -7353,6 +7254,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -7377,6 +7280,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -7390,6 +7294,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -7398,6 +7304,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -7414,6 +7321,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -7464,6 +7373,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -7478,12 +7389,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -7497,6 +7411,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -7517,6 +7432,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -7548,6 +7464,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -7557,6 +7474,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -7568,6 +7486,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -7583,6 +7504,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -7608,6 +7531,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -7628,6 +7552,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -7638,12 +7563,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -7659,6 +7587,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -7693,6 +7622,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -7725,217 +7655,9 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 -PREHOOK: query: explain select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n3 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n3 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n3 - filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n3 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n3 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -2 val_2 -4 val_4 5 val_5 5 val_5 5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 51 val_51 51 val_51 53 val_53 @@ -7959,6 +7681,7 @@ POSTHOOK: Input: default@src 76 val_76 77 val_77 78 val_78 +8 val_8 80 val_80 82 val_82 83 val_83 @@ -7968,6 +7691,7 @@ POSTHOOK: Input: default@src 85 val_85 86 val_86 87 val_87 +9 val_9 90 val_90 90 val_90 90 val_90 @@ -7979,6 +7703,179 @@ POSTHOOK: Input: default@src 97 val_97 98 val_98 98 val_98 +PREHOOK: query: explain select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n3 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n3 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n3 + filterExpr: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key > 0) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n3 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_subq_n3 where masking_test_subq_n3.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n3 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 100 val_100 100 val_100 103 val_103 @@ -7986,6 +7883,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -7996,6 +7894,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -8023,6 +7923,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -8047,6 +7949,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -8060,6 +7963,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -8068,6 +7973,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -8084,6 +7990,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -8134,6 +8042,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -8148,12 +8058,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -8167,6 +8080,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -8187,6 +8101,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -8218,6 +8133,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -8227,6 +8143,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -8238,6 +8155,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -8253,6 +8173,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -8278,6 +8200,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -8298,6 +8221,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -8308,12 +8232,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -8329,6 +8256,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -8363,6 +8291,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -8395,6 +8324,54 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +80 val_80 +82 val_82 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +87 val_87 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +98 val_98 +98 val_98 PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -8404,36 +8381,20 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), upper(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) + Filter Operator + predicate: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) + Select Operator + expressions: key (type: string), upper(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink diff --git ql/src/test/results/clientpositive/llap/masking_4.q.out ql/src/test/results/clientpositive/llap/masking_4.q.out new file mode 100644 index 0000000000..8e05859b01 --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_4.q.out @@ -0,0 +1,282 @@ +PREHOOK: query: create table masking_test_n11 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n11 +POSTHOOK: query: create table masking_test_n11 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n11 +POSTHOOK: Lineage: masking_test_n11.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n11.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: create table masking_test_subq_n2 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_subq_n2 +POSTHOOK: query: create table masking_test_subq_n2 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_subq_n2 +POSTHOOK: Lineage: masking_test_subq_n2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_subq_n2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +with q1 as ( select key from q2 where key = '5'), +q2 as ( select key from src where key = '5') +select * from (select key from q1) a +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select key from q2 where key = '5'), +q2 as ( select key from src where key = '5') +select * from (select key from q1) a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '5') (type: boolean) + Filter Operator + predicate: (key = '5') (type: boolean) + Select Operator + expressions: '5' (type: string) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: explain +with q1 as ( select * from masking_test_n11 where key = '5') +select * from q1 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n11 +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select * from masking_test_n11 where key = '5') +select * from q1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n11 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n11 + filterExpr: ((key = 5) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key = 5) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: 5 (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: explain +with masking_test_subq_n2 as ( select * from masking_test_n11 where key = '5') +select * from masking_test_subq_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n11 +#### A masked pattern was here #### +POSTHOOK: query: explain +with masking_test_subq_n2 as ( select * from masking_test_n11 where key = '5') +select * from masking_test_subq_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n11 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n11 + filterExpr: ((key = 5) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key = 5) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: 5 (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: explain +with q1 as ( select * from masking_test_n11 where key = '5') +select * from masking_test_subq_n2 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n2 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select * from masking_test_n11 where key = '5') +select * from masking_test_subq_n2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n2 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) + Reducer 4 <- Map 3 (SIMPLE_EDGE), Reducer 6 (SIMPLE_EDGE) + Reducer 6 <- Map 5 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n2 + filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: masking_test_subq_n2 + filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.5 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double), _col0 (type: int) + 1 UDFToDouble(_col0) (type: double), _col1 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: double) + 1 _col1 (type: double) + outputColumnNames: _col0, _col2 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: int) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) + Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE + Reducer 6 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: double) + Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/masking_6.q.out ql/src/test/results/clientpositive/llap/masking_6.q.out similarity index 50% rename from ql/src/test/results/clientpositive/masking_6.q.out rename to ql/src/test/results/clientpositive/llap/masking_6.q.out index 1c486d2840..f05cb846e1 100644 --- ql/src/test/results/clientpositive/masking_6.q.out +++ ql/src/test/results/clientpositive/llap/masking_6.q.out @@ -26,40 +26,24 @@ POSTHOOK: Input: default@masking_test_n0 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 83 Data size: 22742 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 22742 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2 + ListSink PREHOOK: query: select * from masking_test_n0 PREHOOK: type: QUERY @@ -88,40 +72,24 @@ POSTHOOK: Input: default@masking_test_n0 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 7398 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 7398 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2 + ListSink PREHOOK: query: select * from masking_test_n0 where key > 0 PREHOOK: type: QUERY @@ -181,40 +149,24 @@ POSTHOOK: Input: default@masking_test_n0 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 7221 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 83 Data size: 78850 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 78850 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + ListSink PREHOOK: query: select * from masking_test_n0 PREHOOK: type: QUERY @@ -243,40 +195,24 @@ POSTHOOK: Input: default@masking_test_n0 POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 2349 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 27 Data size: 25650 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 25650 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + ListSink PREHOOK: query: select * from masking_test_n0 where key > 0 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/masking_7.q.out ql/src/test/results/clientpositive/llap/masking_7.q.out new file mode 100644 index 0000000000..3a6be12241 --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_7.q.out @@ -0,0 +1,231 @@ +PREHOOK: query: drop view masking_test_n14 +PREHOOK: type: DROPVIEW +POSTHOOK: query: drop view masking_test_n14 +POSTHOOK: type: DROPVIEW +PREHOOK: query: create view masking_test_n14 as select cast(key as int) as key, value, '12', ROW__ID from src +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n14 +POSTHOOK: query: create view masking_test_n14 as select cast(key as int) as key, value, '12', ROW__ID from src +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n14 +POSTHOOK: Lineage: masking_test_n14._c2 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n14.row__id SIMPLE [(src)src.FieldSchema(name:ROW__ID, type:struct, comment:), ] +POSTHOOK: Lineage: masking_test_n14.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select * from masking_test_n14 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n14 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string), ROW__ID (type: struct) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink + +PREHOOK: query: select * from masking_test_n14 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n14 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 0_lav 12 NULL +4 4_lav 12 NULL +8 8_lav 12 NULL +0 0_lav 12 NULL +0 0_lav 12 NULL +2 2_lav 12 NULL +PREHOOK: query: explain select * from masking_test_n14 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n14 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string), ROW__ID (type: struct) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink + +PREHOOK: query: select * from masking_test_n14 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n14 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +4 4_lav 12 NULL +8 8_lav 12 NULL +2 2_lav 12 NULL +PREHOOK: query: drop view masking_test_n14 +PREHOOK: type: DROPVIEW +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Output: default@masking_test_n14 +POSTHOOK: query: drop view masking_test_n14 +POSTHOOK: type: DROPVIEW +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Output: default@masking_test_n14 +PREHOOK: query: create view masking_test_n14 as select cast(key as int) as key, '12', ROW__ID, +'12', '12', '12', '12', '12', '12', '12', '12', '12', '12' + from src +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n14 +POSTHOOK: query: create view masking_test_n14 as select cast(key as int) as key, '12', ROW__ID, +'12', '12', '12', '12', '12', '12', '12', '12', '12', '12' + from src +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n14 +POSTHOOK: Lineage: masking_test_n14._c1 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c10 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c11 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c12 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c3 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c4 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c5 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c6 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c7 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c8 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14._c9 SIMPLE [] +POSTHOOK: Lineage: masking_test_n14.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n14.row__id SIMPLE [(src)src.FieldSchema(name:ROW__ID, type:struct, comment:), ] +PREHOOK: query: explain select * from masking_test_n14 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n14 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), '12' (type: string), ROW__ID (type: struct), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + ListSink + +PREHOOK: query: select * from masking_test_n14 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n14 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 12 NULL 12 12 12 12 12 12 12 12 12 12 +4 12 NULL 12 12 12 12 12 12 12 12 12 12 +8 12 NULL 12 12 12 12 12 12 12 12 12 12 +0 12 NULL 12 12 12 12 12 12 12 12 12 12 +0 12 NULL 12 12 12 12 12 12 12 12 12 12 +2 12 NULL 12 12 12 12 12 12 12 12 12 12 +PREHOOK: query: explain select * from masking_test_n14 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n14 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + properties: + insideView TRUE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), '12' (type: string), ROW__ID (type: struct), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 + ListSink + +PREHOOK: query: select * from masking_test_n14 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n14 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n14 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n14 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +4 12 NULL 12 12 12 12 12 12 12 12 12 12 +8 12 NULL 12 12 12 12 12 12 12 12 12 12 +2 12 NULL 12 12 12 12 12 12 12 12 12 12 diff --git ql/src/test/results/clientpositive/masking_8.q.out ql/src/test/results/clientpositive/llap/masking_8.q.out similarity index 57% rename from ql/src/test/results/clientpositive/masking_8.q.out rename to ql/src/test/results/clientpositive/llap/masking_8.q.out index ee11d376ad..c29be1c975 100644 --- ql/src/test/results/clientpositive/masking_8.q.out +++ ql/src/test/results/clientpositive/llap/masking_8.q.out @@ -24,38 +24,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n2 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n2 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 90500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 905 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string), _c2 (type: string), ROW__ID (type: struct) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 5 Data size: 1750 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 1750 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n2 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string), _c2 (type: string), ROW__ID (type: struct) + outputColumnNames: _col0, _col1, _col2, _col3 + ListSink PREHOOK: query: select *, ROW__ID from masking_test_n2 PREHOOK: type: QUERY @@ -80,38 +64,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n2 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n2 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 90500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 905 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string), _c2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5 Data size: 1370 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 1370 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n2 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string), _c2 (type: string) + outputColumnNames: _col0, _col1, _col2 + ListSink PREHOOK: query: select * from masking_test_n2 PREHOOK: type: QUERY @@ -136,37 +104,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n2 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n2 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 90500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 905 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: INPUT__FILE__NAME (type: string), key (type: int), reverse(value) (type: string), _c2 (type: string), ROW__ID (type: struct) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 5 Data size: 2670 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 2670 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n2 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: INPUT__FILE__NAME (type: string), key (type: int), reverse(value) (type: string), _c2 (type: string), ROW__ID (type: struct) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + ListSink PREHOOK: query: select INPUT__FILE__NAME, *, ROW__ID from masking_test_n2 PREHOOK: type: QUERY @@ -218,38 +171,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n2 #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n2 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 433000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 4330 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: ROW__ID (type: struct), key (type: int), _c1 (type: string), _c2 (type: string), _c3 (type: string), _c4 (type: string), _c5 (type: string), _c6 (type: string), _c7 (type: string), _c8 (type: string), _c9 (type: string), _c10 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 - Statistics: Num rows: 5 Data size: 4710 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 4710 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test_n2 + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: ROW__ID (type: struct), key (type: int), _c1 (type: string), _c2 (type: string), _c3 (type: string), _c4 (type: string), _c5 (type: string), _c6 (type: string), _c7 (type: string), _c8 (type: string), _c9 (type: string), _c10 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11 + ListSink PREHOOK: query: select ROW__ID, * from masking_test_n2 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/masking_9.q.out ql/src/test/results/clientpositive/llap/masking_9.q.out similarity index 50% rename from ql/src/test/results/clientpositive/masking_9.q.out rename to ql/src/test/results/clientpositive/llap/masking_9.q.out index 1a71236034..6ec13475b4 100644 --- ql/src/test/results/clientpositive/masking_9.q.out +++ ql/src/test/results/clientpositive/llap/masking_9.q.out @@ -21,38 +21,22 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 1904 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 250 Data size: 952 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: ROW__ID (type: struct) - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 952 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 952 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: masking_test + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: ROW__ID (type: struct) + outputColumnNames: _col0 + ListSink PREHOOK: query: select ROW__ID from masking_test PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap/masking_disablecbo_1.q.out ql/src/test/results/clientpositive/llap/masking_disablecbo_1.q.out new file mode 100644 index 0000000000..8d812b9c9b --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_disablecbo_1.q.out @@ -0,0 +1,428 @@ +PREHOOK: query: create table masking_test_n10 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n10 +POSTHOOK: query: create table masking_test_n10 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n10 +POSTHOOK: Lineage: masking_test_n10.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n10.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select * from masking_test_n10 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n10 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +0 0_lav +0 0_lav +0 0_lav +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select key from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select key from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: select key from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select key from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +2 +4 +8 +PREHOOK: query: explain select value from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select value from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: select value from masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select value from masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +2_lav +4_lav +8_lav +PREHOOK: query: explain select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and UDFToDouble(key) is not null) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 2280 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col0) (type: double) + Statistics: Num rows: 5 Data size: 2280 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(key) (type: double) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double) + 1 UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8 + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +PREHOOK: Input: default@srcpart +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +POSTHOOK: Input: default@srcpart +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 +POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 +#### A masked pattern was here #### +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 11 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-08 12 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 11 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +0 0_lav 0 val_0 2008-04-09 12 +2 2_lav 2 val_2 2008-04-08 11 +2 2_lav 2 val_2 2008-04-08 12 +2 2_lav 2 val_2 2008-04-09 11 +2 2_lav 2 val_2 2008-04-09 12 +4 4_lav 4 val_4 2008-04-08 11 +4 4_lav 4 val_4 2008-04-08 12 +4 4_lav 4 val_4 2008-04-09 11 +4 4_lav 4 val_4 2008-04-09 12 +8 8_lav 8 val_8 2008-04-08 11 +8 8_lav 8 val_8 2008-04-08 12 +8 8_lav 8 val_8 2008-04-09 11 +8 8_lav 8 val_8 2008-04-09 12 +PREHOOK: query: explain select * from default.masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from default.masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from default.masking_test_n10 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select * from default.masking_test_n10 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n10 where masking_test_n10.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n10 where masking_test_n10.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n10 + filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) + Select Operator + expressions: key (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n10 where masking_test_n10.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n10 where masking_test_n10.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n10 +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: string), upper(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + diff --git ql/src/test/results/clientpositive/llap/masking_disablecbo_2.q.out ql/src/test/results/clientpositive/llap/masking_disablecbo_2.q.out new file mode 100644 index 0000000000..41d06040eb --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_disablecbo_2.q.out @@ -0,0 +1,381 @@ +PREHOOK: query: create view masking_test_n13 as select cast(key as int) as key, value from src +PREHOOK: type: CREATEVIEW +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n13 +POSTHOOK: query: create view masking_test_n13 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATEVIEW +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n13 +POSTHOOK: Lineage: masking_test_n13.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n13.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain select * from masking_test_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n13 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n13 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +0 0_lav +0 0_lav +0 0_lav +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from masking_test_n13 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n13 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: select * from masking_test_n13 where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_n13 where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +2 2_lav +4 4_lav +8 8_lav +PREHOOK: query: explain select * from src a join masking_test_n13 b on a.key = b.value where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from src a join masking_test_n13 b on a.key = b.value where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and reverse(value) is not null and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and reverse(value) is not null and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col1 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: string) + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: a + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: key (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: key (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 key (type: string) + 1 _col1 (type: string) + outputColumnNames: _col0, _col1, _col5, _col6 + Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select a.*, b.key from masking_test_n13 a join masking_test_n13 b on a.key = b.value where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select a.*, b.key from masking_test_n13 a join masking_test_n13 b on a.key = b.value where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) + Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col0) (type: double) + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(reverse(value)) is not null and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(reverse(value)) is not null and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col1) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col1) (type: double) + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double) + 1 UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * from masking_test_n13 a union select b.* from masking_test_n13 b where b.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n13 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_n13 a union select b.* from masking_test_n13 b where b.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n13 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 4 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) + Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: int), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) + Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: int), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: int), _col1 (type: string) + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: int), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 2 + Vertex: Union 2 + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/masking_disablecbo_3.q.out ql/src/test/results/clientpositive/llap/masking_disablecbo_3.q.out similarity index 86% rename from ql/src/test/results/clientpositive/masking_disablecbo_3.q.out rename to ql/src/test/results/clientpositive/llap/masking_disablecbo_3.q.out index 587b92791a..c380ebfa1b 100644 --- ql/src/test/results/clientpositive/masking_disablecbo_3.q.out +++ ql/src/test/results/clientpositive/llap/masking_disablecbo_3.q.out @@ -26,61 +26,75 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -101,87 +115,7 @@ POSTHOOK: Input: default@src 0 val_0 0 val_0 0 val_0 -2 val_2 -4 val_4 -5 val_5 -5 val_5 -5 val_5 -8 val_8 -9 val_9 10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -80 val_80 -82 val_82 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -87 val_87 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -98 val_98 -98 val_98 100 val_100 100 val_100 103 val_103 @@ -189,6 +123,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -199,6 +134,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -226,6 +163,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -250,6 +189,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -263,6 +203,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -271,6 +213,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -287,6 +230,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -337,6 +282,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -351,12 +298,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -370,6 +320,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -390,6 +341,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -421,6 +373,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -430,6 +383,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -441,6 +395,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -456,6 +413,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -481,6 +440,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -501,6 +461,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -511,12 +472,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -532,6 +496,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -566,6 +531,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -598,132 +564,9 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 -PREHOOK: query: explain select * from masking_test_subq where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_subq where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_subq where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_subq where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq -POSTHOOK: Input: default@src -#### A masked pattern was here #### -2 val_2 -4 val_4 5 val_5 5 val_5 5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 51 val_51 51 val_51 53 val_53 @@ -747,6 +590,7 @@ POSTHOOK: Input: default@src 76 val_76 77 val_77 78 val_78 +8 val_8 80 val_80 82 val_82 83 val_83 @@ -756,6 +600,7 @@ POSTHOOK: Input: default@src 85 val_85 86 val_86 87 val_87 +9 val_9 90 val_90 90 val_90 90 val_90 @@ -767,6 +612,109 @@ POSTHOOK: Input: default@src 97 val_97 98 val_98 98 val_98 +PREHOOK: query: explain select * from masking_test_subq where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_subq where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_subq where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_subq where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 100 val_100 100 val_100 103 val_103 @@ -774,6 +722,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -784,6 +733,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -811,6 +762,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -835,6 +788,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -848,6 +802,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -856,6 +812,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -872,6 +829,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -922,6 +881,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -936,12 +897,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -955,6 +919,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -975,6 +940,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -1006,6 +972,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -1015,6 +982,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -1026,8 +994,11 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 -351 val_351 -353 val_353 +35 val_35 +35 val_35 +35 val_35 +351 val_351 +353 val_353 353 val_353 356 val_356 360 val_360 @@ -1041,6 +1012,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -1066,6 +1039,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -1086,6 +1060,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -1096,12 +1071,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -1117,6 +1095,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -1151,6 +1130,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -1183,6 +1163,54 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +80 val_80 +82 val_82 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +87 val_87 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +98 val_98 +98 val_98 PREHOOK: query: explain select key from masking_test_subq where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq @@ -1199,61 +1227,75 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0 - Statistics: Num rows: 412 Data size: 1648 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 1648 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0 + Statistics: Num rows: 412 Data size: 1648 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 1648 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -1271,87 +1313,7 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_subq POSTHOOK: Input: default@src #### A masked pattern was here #### -2 -4 -5 -5 -5 -8 -9 10 -11 -12 -12 -15 -15 -17 -18 -18 -19 -20 -24 -24 -26 -26 -27 -28 -30 -33 -34 -35 -35 -35 -37 -37 -41 -42 -42 -43 -44 -47 -51 -51 -53 -54 -57 -58 -58 -64 -65 -66 -67 -67 -69 -70 -70 -70 -72 -72 -74 -76 -76 -77 -78 -80 -82 -83 -83 -84 -84 -85 -86 -87 -90 -90 -90 -92 -95 -95 -96 -97 -97 -98 -98 100 100 103 @@ -1359,6 +1321,7 @@ POSTHOOK: Input: default@src 104 104 105 +11 111 113 113 @@ -1369,6 +1332,8 @@ POSTHOOK: Input: default@src 119 119 119 +12 +12 120 120 125 @@ -1396,6 +1361,8 @@ POSTHOOK: Input: default@src 146 149 149 +15 +15 150 152 152 @@ -1420,6 +1387,7 @@ POSTHOOK: Input: default@src 169 169 169 +17 170 172 172 @@ -1433,6 +1401,8 @@ POSTHOOK: Input: default@src 178 179 179 +18 +18 180 181 183 @@ -1441,6 +1411,7 @@ POSTHOOK: Input: default@src 187 187 189 +19 190 191 191 @@ -1457,6 +1428,8 @@ POSTHOOK: Input: default@src 199 199 199 +2 +20 200 200 201 @@ -1507,6 +1480,8 @@ POSTHOOK: Input: default@src 238 239 239 +24 +24 241 242 242 @@ -1521,12 +1496,15 @@ POSTHOOK: Input: default@src 256 257 258 +26 +26 260 262 263 265 265 266 +27 272 272 273 @@ -1540,6 +1518,7 @@ POSTHOOK: Input: default@src 277 278 278 +28 280 280 281 @@ -1560,6 +1539,7 @@ POSTHOOK: Input: default@src 298 298 298 +30 302 305 306 @@ -1591,6 +1571,7 @@ POSTHOOK: Input: default@src 327 327 327 +33 331 331 332 @@ -1600,6 +1581,7 @@ POSTHOOK: Input: default@src 336 338 339 +34 341 342 342 @@ -1611,6 +1593,9 @@ POSTHOOK: Input: default@src 348 348 348 +35 +35 +35 351 353 353 @@ -1626,6 +1611,8 @@ POSTHOOK: Input: default@src 369 369 369 +37 +37 373 374 375 @@ -1651,6 +1638,7 @@ POSTHOOK: Input: default@src 397 399 399 +4 400 401 401 @@ -1671,6 +1659,7 @@ POSTHOOK: Input: default@src 409 409 409 +41 411 413 413 @@ -1681,12 +1670,15 @@ POSTHOOK: Input: default@src 417 418 419 +42 +42 421 424 424 427 429 429 +43 430 430 430 @@ -1702,6 +1694,7 @@ POSTHOOK: Input: default@src 438 439 439 +44 443 444 446 @@ -1736,6 +1729,7 @@ POSTHOOK: Input: default@src 469 469 469 +47 470 472 475 @@ -1768,6 +1762,54 @@ POSTHOOK: Input: default@src 498 498 498 +5 +5 +5 +51 +51 +53 +54 +57 +58 +58 +64 +65 +66 +67 +67 +69 +70 +70 +70 +72 +72 +74 +76 +76 +77 +78 +8 +80 +82 +83 +83 +84 +84 +85 +86 +87 +9 +90 +90 +90 +92 +95 +95 +96 +97 +97 +98 +98 PREHOOK: query: explain select value from masking_test_subq where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq @@ -1784,65 +1826,79 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col1 - Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col1 + Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 37492 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -1860,94 +1916,15 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_subq POSTHOOK: Input: default@src #### A masked pattern was here #### -val_2 -val_4 -val_5 -val_5 -val_5 -val_8 -val_9 val_10 +val_100 +val_100 +val_103 +val_103 +val_104 +val_104 +val_105 val_11 -val_12 -val_12 -val_15 -val_15 -val_17 -val_18 -val_18 -val_19 -val_20 -val_24 -val_24 -val_26 -val_26 -val_27 -val_28 -val_30 -val_33 -val_34 -val_35 -val_35 -val_35 -val_37 -val_37 -val_41 -val_42 -val_42 -val_43 -val_44 -val_47 -val_51 -val_51 -val_53 -val_54 -val_57 -val_58 -val_58 -val_64 -val_65 -val_66 -val_67 -val_67 -val_69 -val_70 -val_70 -val_70 -val_72 -val_72 -val_74 -val_76 -val_76 -val_77 -val_78 -val_80 -val_82 -val_83 -val_83 -val_84 -val_84 -val_85 -val_86 -val_87 -val_90 -val_90 -val_90 -val_92 -val_95 -val_95 -val_96 -val_97 -val_97 -val_98 -val_98 -val_100 -val_100 -val_103 -val_103 -val_104 -val_104 -val_105 val_111 val_113 val_113 @@ -1958,6 +1935,8 @@ val_118 val_119 val_119 val_119 +val_12 +val_12 val_120 val_120 val_125 @@ -1985,6 +1964,8 @@ val_146 val_146 val_149 val_149 +val_15 +val_15 val_150 val_152 val_152 @@ -2009,6 +1990,7 @@ val_169 val_169 val_169 val_169 +val_17 val_170 val_172 val_172 @@ -2022,6 +2004,8 @@ val_177 val_178 val_179 val_179 +val_18 +val_18 val_180 val_181 val_183 @@ -2030,6 +2014,7 @@ val_187 val_187 val_187 val_189 +val_19 val_190 val_191 val_191 @@ -2046,6 +2031,8 @@ val_197 val_199 val_199 val_199 +val_2 +val_20 val_200 val_200 val_201 @@ -2096,6 +2083,8 @@ val_238 val_238 val_239 val_239 +val_24 +val_24 val_241 val_242 val_242 @@ -2110,12 +2099,15 @@ val_256 val_256 val_257 val_258 +val_26 +val_26 val_260 val_262 val_263 val_265 val_265 val_266 +val_27 val_272 val_272 val_273 @@ -2129,6 +2121,7 @@ val_277 val_277 val_278 val_278 +val_28 val_280 val_280 val_281 @@ -2149,6 +2142,7 @@ val_296 val_298 val_298 val_298 +val_30 val_302 val_305 val_306 @@ -2180,6 +2174,7 @@ val_325 val_327 val_327 val_327 +val_33 val_331 val_331 val_332 @@ -2189,6 +2184,7 @@ val_335 val_336 val_338 val_339 +val_34 val_341 val_342 val_342 @@ -2200,6 +2196,9 @@ val_348 val_348 val_348 val_348 +val_35 +val_35 +val_35 val_351 val_353 val_353 @@ -2215,6 +2214,8 @@ val_368 val_369 val_369 val_369 +val_37 +val_37 val_373 val_374 val_375 @@ -2240,6 +2241,7 @@ val_397 val_397 val_399 val_399 +val_4 val_400 val_401 val_401 @@ -2260,6 +2262,7 @@ val_407 val_409 val_409 val_409 +val_41 val_411 val_413 val_413 @@ -2270,12 +2273,15 @@ val_417 val_417 val_418 val_419 +val_42 +val_42 val_421 val_424 val_424 val_427 val_429 val_429 +val_43 val_430 val_430 val_430 @@ -2291,6 +2297,7 @@ val_438 val_438 val_439 val_439 +val_44 val_443 val_444 val_446 @@ -2325,6 +2332,7 @@ val_469 val_469 val_469 val_469 +val_47 val_470 val_472 val_475 @@ -2357,6 +2365,54 @@ val_497 val_498 val_498 val_498 +val_5 +val_5 +val_5 +val_51 +val_51 +val_53 +val_54 +val_57 +val_58 +val_58 +val_64 +val_65 +val_66 +val_67 +val_67 +val_69 +val_70 +val_70 +val_70 +val_72 +val_72 +val_74 +val_76 +val_76 +val_77 +val_78 +val_8 +val_80 +val_82 +val_83 +val_83 +val_84 +val_84 +val_85 +val_86 +val_87 +val_9 +val_90 +val_90 +val_90 +val_92 +val_95 +val_95 +val_96 +val_97 +val_97 +val_98 +val_98 PREHOOK: query: explain select * from masking_test_subq join srcpart on (masking_test_subq.key = srcpart.key) PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq @@ -2379,111 +2435,120 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 #### A masked pattern was here #### STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 5 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(key) (type: double) + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col0) (type: double) - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - alias: srcpart - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(key) (type: double) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double) - 1 UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8 - Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double) + null sort order: z + sort order: + + Map-reduce partition columns: UDFToDouble(_col0) (type: double) + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: int), _col1 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 UDFToDouble(_col0) (type: double) + 1 UDFToDouble(key) (type: double) + outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8 + Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col0 (type: int), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 2607 Data size: 1671087 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -2511,4118 +2576,4118 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 #### A masked pattern was here #### -0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-09 11 -0 val_0 0 val_0 2008-04-09 12 -0 val_0 0 val_0 2008-04-09 12 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-08 11 +0 val_0 0 val_0 2008-04-08 12 +0 val_0 0 val_0 2008-04-08 12 +0 val_0 0 val_0 2008-04-08 12 +0 val_0 0 val_0 2008-04-08 12 +0 val_0 0 val_0 2008-04-08 12 +0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-08 12 -0 val_0 0 val_0 2008-04-08 11 -0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-08 12 0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 +0 val_0 0 val_0 2008-04-09 11 0 val_0 0 val_0 2008-04-09 12 -2 val_2 2 val_2 2008-04-09 11 -2 val_2 2 val_2 2008-04-08 11 -2 val_2 2 val_2 2008-04-09 12 -2 val_2 2 val_2 2008-04-08 12 -4 val_4 4 val_4 2008-04-08 12 -4 val_4 4 val_4 2008-04-09 12 -4 val_4 4 val_4 2008-04-08 11 -4 val_4 4 val_4 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-08 11 -5 val_5 5 val_5 2008-04-09 12 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-09 11 -5 val_5 5 val_5 2008-04-08 12 -5 val_5 5 val_5 2008-04-09 12 -8 val_8 8 val_8 2008-04-08 11 -8 val_8 8 val_8 2008-04-09 11 -8 val_8 8 val_8 2008-04-08 12 -8 val_8 8 val_8 2008-04-09 12 -9 val_9 9 val_9 2008-04-08 12 -9 val_9 9 val_9 2008-04-09 12 -9 val_9 9 val_9 2008-04-09 11 -9 val_9 9 val_9 2008-04-08 11 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +0 val_0 0 val_0 2008-04-09 12 +10 val_10 10 val_10 2008-04-08 11 +10 val_10 10 val_10 2008-04-08 12 10 val_10 10 val_10 2008-04-09 11 10 val_10 10 val_10 2008-04-09 12 -10 val_10 10 val_10 2008-04-08 12 -10 val_10 10 val_10 2008-04-08 11 -11 val_11 11 val_11 2008-04-09 12 -11 val_11 11 val_11 2008-04-09 11 -11 val_11 11 val_11 2008-04-08 12 -11 val_11 11 val_11 2008-04-08 11 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-09 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-08 11 -12 val_12 12 val_12 2008-04-08 12 -12 val_12 12 val_12 2008-04-09 11 -12 val_12 12 val_12 2008-04-09 12 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 12 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 12 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 11 -15 val_15 15 val_15 2008-04-09 11 -15 val_15 15 val_15 2008-04-08 12 -17 val_17 17 val_17 2008-04-09 11 -17 val_17 17 val_17 2008-04-09 12 -17 val_17 17 val_17 2008-04-08 12 -17 val_17 17 val_17 2008-04-08 11 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 12 -18 val_18 18 val_18 2008-04-08 11 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 12 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-09 11 -18 val_18 18 val_18 2008-04-08 11 -19 val_19 19 val_19 2008-04-09 11 -19 val_19 19 val_19 2008-04-09 12 -19 val_19 19 val_19 2008-04-08 11 -19 val_19 19 val_19 2008-04-08 12 -20 val_20 20 val_20 2008-04-08 11 -20 val_20 20 val_20 2008-04-09 11 -20 val_20 20 val_20 2008-04-08 12 -20 val_20 20 val_20 2008-04-09 12 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-08 11 -24 val_24 24 val_24 2008-04-09 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-08 12 -24 val_24 24 val_24 2008-04-09 11 -24 val_24 24 val_24 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-08 11 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-09 12 -26 val_26 26 val_26 2008-04-09 11 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 12 -26 val_26 26 val_26 2008-04-08 11 -27 val_27 27 val_27 2008-04-09 12 -27 val_27 27 val_27 2008-04-09 11 -27 val_27 27 val_27 2008-04-08 12 -27 val_27 27 val_27 2008-04-08 11 -28 val_28 28 val_28 2008-04-08 12 -28 val_28 28 val_28 2008-04-09 11 -28 val_28 28 val_28 2008-04-09 12 -28 val_28 28 val_28 2008-04-08 11 -30 val_30 30 val_30 2008-04-09 12 -30 val_30 30 val_30 2008-04-09 11 -30 val_30 30 val_30 2008-04-08 11 -30 val_30 30 val_30 2008-04-08 12 -33 val_33 33 val_33 2008-04-08 11 -33 val_33 33 val_33 2008-04-08 12 -33 val_33 33 val_33 2008-04-09 11 -33 val_33 33 val_33 2008-04-09 12 -34 val_34 34 val_34 2008-04-09 11 -34 val_34 34 val_34 2008-04-09 12 -34 val_34 34 val_34 2008-04-08 11 -34 val_34 34 val_34 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-08 12 -35 val_35 35 val_35 2008-04-09 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-08 11 -35 val_35 35 val_35 2008-04-09 12 -35 val_35 35 val_35 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 11 -37 val_37 37 val_37 2008-04-09 12 -37 val_37 37 val_37 2008-04-08 12 -37 val_37 37 val_37 2008-04-08 11 -41 val_41 41 val_41 2008-04-08 11 -41 val_41 41 val_41 2008-04-09 11 -41 val_41 41 val_41 2008-04-09 12 -41 val_41 41 val_41 2008-04-08 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-08 12 -42 val_42 42 val_42 2008-04-09 12 -42 val_42 42 val_42 2008-04-08 11 -42 val_42 42 val_42 2008-04-09 11 -42 val_42 42 val_42 2008-04-08 11 -43 val_43 43 val_43 2008-04-09 12 -43 val_43 43 val_43 2008-04-08 12 -43 val_43 43 val_43 2008-04-09 11 -43 val_43 43 val_43 2008-04-08 11 -44 val_44 44 val_44 2008-04-09 11 -44 val_44 44 val_44 2008-04-08 11 -44 val_44 44 val_44 2008-04-09 12 -44 val_44 44 val_44 2008-04-08 12 -47 val_47 47 val_47 2008-04-08 11 -47 val_47 47 val_47 2008-04-08 12 -47 val_47 47 val_47 2008-04-09 12 -47 val_47 47 val_47 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-08 11 -51 val_51 51 val_51 2008-04-09 11 -51 val_51 51 val_51 2008-04-09 12 -51 val_51 51 val_51 2008-04-08 12 -51 val_51 51 val_51 2008-04-09 11 -53 val_53 53 val_53 2008-04-08 11 -53 val_53 53 val_53 2008-04-09 11 -53 val_53 53 val_53 2008-04-09 12 -53 val_53 53 val_53 2008-04-08 12 -54 val_54 54 val_54 2008-04-08 11 -54 val_54 54 val_54 2008-04-09 12 -54 val_54 54 val_54 2008-04-09 11 -54 val_54 54 val_54 2008-04-08 12 -57 val_57 57 val_57 2008-04-08 12 -57 val_57 57 val_57 2008-04-08 11 -57 val_57 57 val_57 2008-04-09 11 -57 val_57 57 val_57 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 11 -58 val_58 58 val_58 2008-04-08 12 -58 val_58 58 val_58 2008-04-09 11 -58 val_58 58 val_58 2008-04-09 12 -58 val_58 58 val_58 2008-04-08 12 -64 val_64 64 val_64 2008-04-09 12 -64 val_64 64 val_64 2008-04-08 11 -64 val_64 64 val_64 2008-04-08 12 -64 val_64 64 val_64 2008-04-09 11 -65 val_65 65 val_65 2008-04-09 11 -65 val_65 65 val_65 2008-04-09 12 -65 val_65 65 val_65 2008-04-08 11 -65 val_65 65 val_65 2008-04-08 12 -66 val_66 66 val_66 2008-04-09 12 -66 val_66 66 val_66 2008-04-09 11 -66 val_66 66 val_66 2008-04-08 11 -66 val_66 66 val_66 2008-04-08 12 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-09 12 -67 val_67 67 val_67 2008-04-09 11 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -67 val_67 67 val_67 2008-04-08 11 -67 val_67 67 val_67 2008-04-08 12 -69 val_69 69 val_69 2008-04-09 12 -69 val_69 69 val_69 2008-04-08 11 -69 val_69 69 val_69 2008-04-08 12 -69 val_69 69 val_69 2008-04-09 11 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 12 -70 val_70 70 val_70 2008-04-09 11 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -70 val_70 70 val_70 2008-04-08 11 -70 val_70 70 val_70 2008-04-09 12 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-08 12 -72 val_72 72 val_72 2008-04-09 11 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 12 -72 val_72 72 val_72 2008-04-08 11 -72 val_72 72 val_72 2008-04-09 11 -74 val_74 74 val_74 2008-04-09 11 -74 val_74 74 val_74 2008-04-08 12 -74 val_74 74 val_74 2008-04-09 12 -74 val_74 74 val_74 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 12 -76 val_76 76 val_76 2008-04-08 11 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 12 -76 val_76 76 val_76 2008-04-09 11 -76 val_76 76 val_76 2008-04-09 11 -77 val_77 77 val_77 2008-04-08 12 -77 val_77 77 val_77 2008-04-09 12 -77 val_77 77 val_77 2008-04-09 11 -77 val_77 77 val_77 2008-04-08 11 -78 val_78 78 val_78 2008-04-09 11 -78 val_78 78 val_78 2008-04-08 11 -78 val_78 78 val_78 2008-04-09 12 -78 val_78 78 val_78 2008-04-08 12 -80 val_80 80 val_80 2008-04-09 11 -80 val_80 80 val_80 2008-04-08 11 -80 val_80 80 val_80 2008-04-09 12 -80 val_80 80 val_80 2008-04-08 12 -82 val_82 82 val_82 2008-04-08 12 -82 val_82 82 val_82 2008-04-09 11 -82 val_82 82 val_82 2008-04-08 11 -82 val_82 82 val_82 2008-04-09 12 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-09 12 -83 val_83 83 val_83 2008-04-08 11 -83 val_83 83 val_83 2008-04-08 12 -83 val_83 83 val_83 2008-04-09 11 -83 val_83 83 val_83 2008-04-08 11 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-08 11 -84 val_84 84 val_84 2008-04-09 12 -84 val_84 84 val_84 2008-04-08 12 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-09 11 -84 val_84 84 val_84 2008-04-08 12 -85 val_85 85 val_85 2008-04-09 12 -85 val_85 85 val_85 2008-04-09 11 -85 val_85 85 val_85 2008-04-08 11 -85 val_85 85 val_85 2008-04-08 12 -86 val_86 86 val_86 2008-04-08 12 -86 val_86 86 val_86 2008-04-09 11 -86 val_86 86 val_86 2008-04-08 11 -86 val_86 86 val_86 2008-04-09 12 -87 val_87 87 val_87 2008-04-09 12 -87 val_87 87 val_87 2008-04-08 11 -87 val_87 87 val_87 2008-04-08 12 -87 val_87 87 val_87 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-09 11 -90 val_90 90 val_90 2008-04-09 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 12 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-08 11 -90 val_90 90 val_90 2008-04-09 11 -92 val_92 92 val_92 2008-04-09 11 -92 val_92 92 val_92 2008-04-09 12 -92 val_92 92 val_92 2008-04-08 12 -92 val_92 92 val_92 2008-04-08 11 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-08 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-08 12 -95 val_95 95 val_95 2008-04-09 11 -95 val_95 95 val_95 2008-04-09 12 -95 val_95 95 val_95 2008-04-08 11 -96 val_96 96 val_96 2008-04-09 12 -96 val_96 96 val_96 2008-04-08 11 -96 val_96 96 val_96 2008-04-08 12 -96 val_96 96 val_96 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 12 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-09 11 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 12 -97 val_97 97 val_97 2008-04-08 11 -97 val_97 97 val_97 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-09 11 -98 val_98 98 val_98 2008-04-08 12 -98 val_98 98 val_98 2008-04-09 12 -100 val_100 100 val_100 2008-04-09 11 100 val_100 100 val_100 2008-04-08 11 -100 val_100 100 val_100 2008-04-09 12 +100 val_100 100 val_100 2008-04-08 11 +100 val_100 100 val_100 2008-04-08 11 +100 val_100 100 val_100 2008-04-08 11 100 val_100 100 val_100 2008-04-08 12 100 val_100 100 val_100 2008-04-08 12 -100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-08 11 +100 val_100 100 val_100 2008-04-08 12 +100 val_100 100 val_100 2008-04-08 12 +100 val_100 100 val_100 2008-04-09 11 +100 val_100 100 val_100 2008-04-09 11 100 val_100 100 val_100 2008-04-09 11 100 val_100 100 val_100 2008-04-09 11 -100 val_100 100 val_100 2008-04-08 11 100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-08 12 -100 val_100 100 val_100 2008-04-08 12 100 val_100 100 val_100 2008-04-09 12 -100 val_100 100 val_100 2008-04-08 11 -100 val_100 100 val_100 2008-04-09 11 +100 val_100 100 val_100 2008-04-09 12 +100 val_100 100 val_100 2008-04-09 12 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 12 -103 val_103 103 val_103 2008-04-09 11 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 11 -103 val_103 103 val_103 2008-04-08 12 -103 val_103 103 val_103 2008-04-08 12 -103 val_103 103 val_103 2008-04-09 12 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 12 -103 val_103 103 val_103 2008-04-09 11 103 val_103 103 val_103 2008-04-08 11 -103 val_103 103 val_103 2008-04-09 11 103 val_103 103 val_103 2008-04-08 12 103 val_103 103 val_103 2008-04-08 12 +103 val_103 103 val_103 2008-04-08 12 +103 val_103 103 val_103 2008-04-08 12 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 11 +103 val_103 103 val_103 2008-04-09 12 +103 val_103 103 val_103 2008-04-09 12 +103 val_103 103 val_103 2008-04-09 12 103 val_103 103 val_103 2008-04-09 12 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-09 11 -104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-08 12 -104 val_104 104 val_104 2008-04-09 11 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 -104 val_104 104 val_104 2008-04-09 11 -104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 11 -104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-08 12 +104 val_104 104 val_104 2008-04-08 12 +104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-08 12 104 val_104 104 val_104 2008-04-09 11 -105 val_105 105 val_105 2008-04-09 12 -105 val_105 105 val_105 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 11 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 +104 val_104 104 val_104 2008-04-09 12 105 val_105 105 val_105 2008-04-08 11 105 val_105 105 val_105 2008-04-08 12 -111 val_111 111 val_111 2008-04-08 12 +105 val_105 105 val_105 2008-04-09 11 +105 val_105 105 val_105 2008-04-09 12 +11 val_11 11 val_11 2008-04-08 11 +11 val_11 11 val_11 2008-04-08 12 +11 val_11 11 val_11 2008-04-09 11 +11 val_11 11 val_11 2008-04-09 12 111 val_111 111 val_111 2008-04-08 11 -111 val_111 111 val_111 2008-04-09 12 +111 val_111 111 val_111 2008-04-08 12 111 val_111 111 val_111 2008-04-09 11 -113 val_113 113 val_113 2008-04-09 12 +111 val_111 111 val_111 2008-04-09 12 +113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-08 11 -113 val_113 113 val_113 2008-04-08 12 -113 val_113 113 val_113 2008-04-08 12 -113 val_113 113 val_113 2008-04-09 11 -113 val_113 113 val_113 2008-04-09 11 113 val_113 113 val_113 2008-04-08 11 -113 val_113 113 val_113 2008-04-09 12 -113 val_113 113 val_113 2008-04-09 12 113 val_113 113 val_113 2008-04-08 11 113 val_113 113 val_113 2008-04-08 12 113 val_113 113 val_113 2008-04-08 12 +113 val_113 113 val_113 2008-04-08 12 +113 val_113 113 val_113 2008-04-08 12 113 val_113 113 val_113 2008-04-09 11 113 val_113 113 val_113 2008-04-09 11 -113 val_113 113 val_113 2008-04-08 11 +113 val_113 113 val_113 2008-04-09 11 +113 val_113 113 val_113 2008-04-09 11 +113 val_113 113 val_113 2008-04-09 12 113 val_113 113 val_113 2008-04-09 12 +113 val_113 113 val_113 2008-04-09 12 +113 val_113 113 val_113 2008-04-09 12 +114 val_114 114 val_114 2008-04-08 11 +114 val_114 114 val_114 2008-04-08 12 114 val_114 114 val_114 2008-04-09 11 114 val_114 114 val_114 2008-04-09 12 -114 val_114 114 val_114 2008-04-08 12 -114 val_114 114 val_114 2008-04-08 11 +116 val_116 116 val_116 2008-04-08 11 116 val_116 116 val_116 2008-04-08 12 116 val_116 116 val_116 2008-04-09 11 -116 val_116 116 val_116 2008-04-08 11 116 val_116 116 val_116 2008-04-09 12 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 -118 val_118 118 val_118 2008-04-09 11 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 -118 val_118 118 val_118 2008-04-08 12 -118 val_118 118 val_118 2008-04-08 12 -118 val_118 118 val_118 2008-04-09 11 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 -118 val_118 118 val_118 2008-04-09 11 -118 val_118 118 val_118 2008-04-09 12 118 val_118 118 val_118 2008-04-08 11 118 val_118 118 val_118 2008-04-08 12 118 val_118 118 val_118 2008-04-08 12 +118 val_118 118 val_118 2008-04-08 12 +118 val_118 118 val_118 2008-04-08 12 118 val_118 118 val_118 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 11 +118 val_118 118 val_118 2008-04-09 11 +118 val_118 118 val_118 2008-04-09 11 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +118 val_118 118 val_118 2008-04-09 12 +119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 11 -119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-09 12 119 val_119 119 val_119 2008-04-08 11 -119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 119 val_119 119 val_119 2008-04-08 12 -119 val_119 119 val_119 2008-04-08 11 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-08 12 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 11 119 val_119 119 val_119 2008-04-09 12 -119 val_119 119 val_119 2008-04-08 11 119 val_119 119 val_119 2008-04-09 12 -120 val_120 120 val_120 2008-04-09 11 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +119 val_119 119 val_119 2008-04-09 12 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 11 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-08 12 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 11 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +12 val_12 12 val_12 2008-04-09 12 +120 val_120 120 val_120 2008-04-08 11 +120 val_120 120 val_120 2008-04-08 11 +120 val_120 120 val_120 2008-04-08 11 120 val_120 120 val_120 2008-04-08 11 -120 val_120 120 val_120 2008-04-09 12 120 val_120 120 val_120 2008-04-08 12 -120 val_120 120 val_120 2008-04-09 12 120 val_120 120 val_120 2008-04-08 12 -120 val_120 120 val_120 2008-04-08 11 +120 val_120 120 val_120 2008-04-08 12 +120 val_120 120 val_120 2008-04-08 12 +120 val_120 120 val_120 2008-04-09 11 +120 val_120 120 val_120 2008-04-09 11 120 val_120 120 val_120 2008-04-09 11 120 val_120 120 val_120 2008-04-09 11 -120 val_120 120 val_120 2008-04-08 11 120 val_120 120 val_120 2008-04-09 12 -120 val_120 120 val_120 2008-04-08 12 120 val_120 120 val_120 2008-04-09 12 -120 val_120 120 val_120 2008-04-08 12 -120 val_120 120 val_120 2008-04-08 11 -120 val_120 120 val_120 2008-04-09 11 -125 val_125 125 val_125 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 12 +120 val_120 120 val_120 2008-04-09 12 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 11 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 12 -125 val_125 125 val_125 2008-04-09 11 -125 val_125 125 val_125 2008-04-08 12 -125 val_125 125 val_125 2008-04-08 12 -125 val_125 125 val_125 2008-04-09 12 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 11 125 val_125 125 val_125 2008-04-08 11 -125 val_125 125 val_125 2008-04-09 12 -125 val_125 125 val_125 2008-04-09 11 125 val_125 125 val_125 2008-04-08 12 125 val_125 125 val_125 2008-04-08 12 +125 val_125 125 val_125 2008-04-08 12 +125 val_125 125 val_125 2008-04-08 12 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 11 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 +125 val_125 125 val_125 2008-04-09 12 126 val_126 126 val_126 2008-04-08 11 126 val_126 126 val_126 2008-04-08 12 -126 val_126 126 val_126 2008-04-09 12 126 val_126 126 val_126 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 11 +126 val_126 126 val_126 2008-04-09 12 128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 128 val_128 128 val_128 2008-04-08 11 -128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-08 11 128 val_128 128 val_128 2008-04-08 11 +128 val_128 128 val_128 2008-04-08 11 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 +128 val_128 128 val_128 2008-04-08 12 128 val_128 128 val_128 2008-04-08 12 128 val_128 128 val_128 2008-04-09 11 -128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 11 128 val_128 128 val_128 2008-04-09 12 -128 val_128 128 val_128 2008-04-08 12 -128 val_128 128 val_128 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 12 -129 val_129 129 val_129 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 12 -129 val_129 129 val_129 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 11 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 +128 val_128 128 val_128 2008-04-09 12 129 val_129 129 val_129 2008-04-08 11 -129 val_129 129 val_129 2008-04-09 11 129 val_129 129 val_129 2008-04-08 11 -129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-08 12 +129 val_129 129 val_129 2008-04-08 12 129 val_129 129 val_129 2008-04-08 12 -129 val_129 129 val_129 2008-04-09 12 129 val_129 129 val_129 2008-04-08 12 129 val_129 129 val_129 2008-04-09 11 -129 val_129 129 val_129 2008-04-08 11 129 val_129 129 val_129 2008-04-09 11 -129 val_129 129 val_129 2008-04-08 11 +129 val_129 129 val_129 2008-04-09 11 +129 val_129 129 val_129 2008-04-09 11 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +129 val_129 129 val_129 2008-04-09 12 +131 val_131 131 val_131 2008-04-08 11 131 val_131 131 val_131 2008-04-08 12 -131 val_131 131 val_131 2008-04-09 12 131 val_131 131 val_131 2008-04-09 11 -131 val_131 131 val_131 2008-04-08 11 +131 val_131 131 val_131 2008-04-09 12 133 val_133 133 val_133 2008-04-08 11 133 val_133 133 val_133 2008-04-08 12 133 val_133 133 val_133 2008-04-09 11 133 val_133 133 val_133 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 12 -134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-08 11 -134 val_134 134 val_134 2008-04-09 11 -134 val_134 134 val_134 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 12 -134 val_134 134 val_134 2008-04-09 12 134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-08 12 +134 val_134 134 val_134 2008-04-08 12 +134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-09 11 -134 val_134 134 val_134 2008-04-08 11 +134 val_134 134 val_134 2008-04-09 11 +134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-09 11 134 val_134 134 val_134 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 12 134 val_134 134 val_134 2008-04-09 12 -134 val_134 134 val_134 2008-04-08 11 -136 val_136 136 val_136 2008-04-09 12 -136 val_136 136 val_136 2008-04-08 12 +134 val_134 134 val_134 2008-04-09 12 +134 val_134 134 val_134 2008-04-09 12 136 val_136 136 val_136 2008-04-08 11 +136 val_136 136 val_136 2008-04-08 12 136 val_136 136 val_136 2008-04-09 11 +136 val_136 136 val_136 2008-04-09 12 137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-09 12 -137 val_137 137 val_137 2008-04-09 11 -137 val_137 137 val_137 2008-04-09 11 -137 val_137 137 val_137 2008-04-09 12 -137 val_137 137 val_137 2008-04-08 12 137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-08 12 137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-09 12 +137 val_137 137 val_137 2008-04-08 11 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-08 12 +137 val_137 137 val_137 2008-04-09 11 +137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 11 137 val_137 137 val_137 2008-04-09 12 -137 val_137 137 val_137 2008-04-08 12 -137 val_137 137 val_137 2008-04-08 11 -137 val_137 137 val_137 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 +137 val_137 137 val_137 2008-04-09 12 +137 val_137 137 val_137 2008-04-09 12 +137 val_137 137 val_137 2008-04-09 12 +138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 -138 val_138 138 val_138 2008-04-09 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-08 12 138 val_138 138 val_138 2008-04-08 11 -138 val_138 138 val_138 2008-04-09 12 138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 138 val_138 138 val_138 2008-04-08 12 -138 val_138 138 val_138 2008-04-09 11 -138 val_138 138 val_138 2008-04-08 11 138 val_138 138 val_138 2008-04-08 12 -143 val_143 143 val_143 2008-04-09 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-08 12 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 11 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 +138 val_138 138 val_138 2008-04-09 12 143 val_143 143 val_143 2008-04-08 11 -143 val_143 143 val_143 2008-04-09 11 143 val_143 143 val_143 2008-04-08 12 -145 val_145 145 val_145 2008-04-09 11 +143 val_143 143 val_143 2008-04-09 11 +143 val_143 143 val_143 2008-04-09 12 145 val_145 145 val_145 2008-04-08 11 145 val_145 145 val_145 2008-04-08 12 +145 val_145 145 val_145 2008-04-09 11 145 val_145 145 val_145 2008-04-09 12 146 val_146 146 val_146 2008-04-08 11 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 146 val_146 146 val_146 2008-04-08 11 -146 val_146 146 val_146 2008-04-09 12 -146 val_146 146 val_146 2008-04-09 12 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 146 val_146 146 val_146 2008-04-08 11 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 146 val_146 146 val_146 2008-04-08 11 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-08 12 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 11 146 val_146 146 val_146 2008-04-09 12 146 val_146 146 val_146 2008-04-09 12 -146 val_146 146 val_146 2008-04-09 11 -146 val_146 146 val_146 2008-04-08 12 -149 val_149 149 val_149 2008-04-08 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 11 -149 val_149 149 val_149 2008-04-09 11 +146 val_146 146 val_146 2008-04-09 12 +146 val_146 146 val_146 2008-04-09 12 +149 val_149 149 val_149 2008-04-08 11 +149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 11 149 val_149 149 val_149 2008-04-08 12 149 val_149 149 val_149 2008-04-08 12 -149 val_149 149 val_149 2008-04-09 12 -149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-08 12 +149 val_149 149 val_149 2008-04-08 12 149 val_149 149 val_149 2008-04-09 11 149 val_149 149 val_149 2008-04-09 11 -149 val_149 149 val_149 2008-04-08 11 -149 val_149 149 val_149 2008-04-08 11 -149 val_149 149 val_149 2008-04-08 12 -150 val_150 150 val_150 2008-04-09 11 -150 val_150 150 val_150 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 11 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +149 val_149 149 val_149 2008-04-09 12 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 11 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-08 12 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 11 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 +15 val_15 15 val_15 2008-04-09 12 150 val_150 150 val_150 2008-04-08 11 150 val_150 150 val_150 2008-04-08 12 +150 val_150 150 val_150 2008-04-09 11 +150 val_150 150 val_150 2008-04-09 12 152 val_152 152 val_152 2008-04-08 11 -152 val_152 152 val_152 2008-04-08 12 -152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-08 11 -152 val_152 152 val_152 2008-04-09 12 -152 val_152 152 val_152 2008-04-09 11 -152 val_152 152 val_152 2008-04-09 12 -152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-08 11 +152 val_152 152 val_152 2008-04-08 11 +152 val_152 152 val_152 2008-04-08 12 +152 val_152 152 val_152 2008-04-08 12 +152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-08 12 152 val_152 152 val_152 2008-04-09 11 -152 val_152 152 val_152 2008-04-08 11 -152 val_152 152 val_152 2008-04-09 12 +152 val_152 152 val_152 2008-04-09 11 +152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-09 11 152 val_152 152 val_152 2008-04-09 12 -152 val_152 152 val_152 2008-04-08 12 -153 val_153 153 val_153 2008-04-09 11 +152 val_152 152 val_152 2008-04-09 12 +152 val_152 152 val_152 2008-04-09 12 +152 val_152 152 val_152 2008-04-09 12 153 val_153 153 val_153 2008-04-08 11 153 val_153 153 val_153 2008-04-08 12 +153 val_153 153 val_153 2008-04-09 11 153 val_153 153 val_153 2008-04-09 12 -155 val_155 155 val_155 2008-04-08 12 155 val_155 155 val_155 2008-04-08 11 -155 val_155 155 val_155 2008-04-09 12 +155 val_155 155 val_155 2008-04-08 12 155 val_155 155 val_155 2008-04-09 11 +155 val_155 155 val_155 2008-04-09 12 +156 val_156 156 val_156 2008-04-08 11 156 val_156 156 val_156 2008-04-08 12 -156 val_156 156 val_156 2008-04-09 12 156 val_156 156 val_156 2008-04-09 11 -156 val_156 156 val_156 2008-04-08 11 -157 val_157 157 val_157 2008-04-09 11 -157 val_157 157 val_157 2008-04-08 12 +156 val_156 156 val_156 2008-04-09 12 157 val_157 157 val_157 2008-04-08 11 +157 val_157 157 val_157 2008-04-08 12 +157 val_157 157 val_157 2008-04-09 11 157 val_157 157 val_157 2008-04-09 12 158 val_158 158 val_158 2008-04-08 11 +158 val_158 158 val_158 2008-04-08 12 158 val_158 158 val_158 2008-04-09 11 158 val_158 158 val_158 2008-04-09 12 -158 val_158 158 val_158 2008-04-08 12 -160 val_160 160 val_160 2008-04-09 11 -160 val_160 160 val_160 2008-04-09 12 160 val_160 160 val_160 2008-04-08 11 160 val_160 160 val_160 2008-04-08 12 -162 val_162 162 val_162 2008-04-09 11 -162 val_162 162 val_162 2008-04-08 12 +160 val_160 160 val_160 2008-04-09 11 +160 val_160 160 val_160 2008-04-09 12 162 val_162 162 val_162 2008-04-08 11 +162 val_162 162 val_162 2008-04-08 12 +162 val_162 162 val_162 2008-04-09 11 162 val_162 162 val_162 2008-04-09 12 -163 val_163 163 val_163 2008-04-09 11 -163 val_163 163 val_163 2008-04-09 12 163 val_163 163 val_163 2008-04-08 11 163 val_163 163 val_163 2008-04-08 12 +163 val_163 163 val_163 2008-04-09 11 +163 val_163 163 val_163 2008-04-09 12 +164 val_164 164 val_164 2008-04-08 11 +164 val_164 164 val_164 2008-04-08 11 164 val_164 164 val_164 2008-04-08 11 -164 val_164 164 val_164 2008-04-09 11 164 val_164 164 val_164 2008-04-08 11 -164 val_164 164 val_164 2008-04-09 12 -164 val_164 164 val_164 2008-04-09 11 -164 val_164 164 val_164 2008-04-09 12 164 val_164 164 val_164 2008-04-08 12 164 val_164 164 val_164 2008-04-08 12 -164 val_164 164 val_164 2008-04-08 11 +164 val_164 164 val_164 2008-04-08 12 +164 val_164 164 val_164 2008-04-08 12 +164 val_164 164 val_164 2008-04-09 11 +164 val_164 164 val_164 2008-04-09 11 164 val_164 164 val_164 2008-04-09 11 -164 val_164 164 val_164 2008-04-08 11 -164 val_164 164 val_164 2008-04-09 12 164 val_164 164 val_164 2008-04-09 11 164 val_164 164 val_164 2008-04-09 12 -164 val_164 164 val_164 2008-04-08 12 -164 val_164 164 val_164 2008-04-08 12 +164 val_164 164 val_164 2008-04-09 12 +164 val_164 164 val_164 2008-04-09 12 +164 val_164 164 val_164 2008-04-09 12 +165 val_165 165 val_165 2008-04-08 11 +165 val_165 165 val_165 2008-04-08 11 +165 val_165 165 val_165 2008-04-08 11 165 val_165 165 val_165 2008-04-08 11 -165 val_165 165 val_165 2008-04-09 11 -165 val_165 165 val_165 2008-04-09 12 165 val_165 165 val_165 2008-04-08 12 -165 val_165 165 val_165 2008-04-09 12 165 val_165 165 val_165 2008-04-08 12 -165 val_165 165 val_165 2008-04-08 11 +165 val_165 165 val_165 2008-04-08 12 +165 val_165 165 val_165 2008-04-08 12 +165 val_165 165 val_165 2008-04-09 11 +165 val_165 165 val_165 2008-04-09 11 165 val_165 165 val_165 2008-04-09 11 -165 val_165 165 val_165 2008-04-08 11 165 val_165 165 val_165 2008-04-09 11 165 val_165 165 val_165 2008-04-09 12 -165 val_165 165 val_165 2008-04-08 12 165 val_165 165 val_165 2008-04-09 12 -165 val_165 165 val_165 2008-04-08 12 -165 val_165 165 val_165 2008-04-08 11 -165 val_165 165 val_165 2008-04-09 11 -166 val_166 166 val_166 2008-04-08 12 +165 val_165 165 val_165 2008-04-09 12 +165 val_165 165 val_165 2008-04-09 12 166 val_166 166 val_166 2008-04-08 11 -166 val_166 166 val_166 2008-04-09 12 +166 val_166 166 val_166 2008-04-08 12 166 val_166 166 val_166 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 11 +166 val_166 166 val_166 2008-04-09 12 +167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 12 -167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-08 11 -167 val_167 167 val_167 2008-04-09 11 -167 val_167 167 val_167 2008-04-08 12 167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 +167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 11 167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 11 167 val_167 167 val_167 2008-04-09 12 167 val_167 167 val_167 2008-04-09 12 -167 val_167 167 val_167 2008-04-08 12 +167 val_167 167 val_167 2008-04-09 12 +167 val_167 167 val_167 2008-04-09 12 +167 val_167 167 val_167 2008-04-09 12 +167 val_167 167 val_167 2008-04-09 12 +167 val_167 167 val_167 2008-04-09 12 +167 val_167 167 val_167 2008-04-09 12 168 val_168 168 val_168 2008-04-08 11 168 val_168 168 val_168 2008-04-08 12 168 val_168 168 val_168 2008-04-09 11 168 val_168 168 val_168 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 12 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 12 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-09 12 169 val_169 169 val_169 2008-04-08 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 -169 val_169 169 val_169 2008-04-08 12 -169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 11 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-08 12 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 +169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-09 11 169 val_169 169 val_169 2008-04-09 12 -169 val_169 169 val_169 2008-04-08 11 -169 val_169 169 val_169 2008-04-08 11 -170 val_170 170 val_170 2008-04-09 11 -170 val_170 170 val_170 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +169 val_169 169 val_169 2008-04-09 12 +17 val_17 17 val_17 2008-04-08 11 +17 val_17 17 val_17 2008-04-08 12 +17 val_17 17 val_17 2008-04-09 11 +17 val_17 17 val_17 2008-04-09 12 170 val_170 170 val_170 2008-04-08 11 170 val_170 170 val_170 2008-04-08 12 +170 val_170 170 val_170 2008-04-09 11 +170 val_170 170 val_170 2008-04-09 12 +172 val_172 172 val_172 2008-04-08 11 +172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-08 11 -172 val_172 172 val_172 2008-04-09 11 -172 val_172 172 val_172 2008-04-08 12 -172 val_172 172 val_172 2008-04-09 12 172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-08 12 -172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-08 12 +172 val_172 172 val_172 2008-04-08 12 +172 val_172 172 val_172 2008-04-08 12 +172 val_172 172 val_172 2008-04-09 11 +172 val_172 172 val_172 2008-04-09 11 172 val_172 172 val_172 2008-04-09 11 -172 val_172 172 val_172 2008-04-08 11 172 val_172 172 val_172 2008-04-09 11 -172 val_172 172 val_172 2008-04-08 12 172 val_172 172 val_172 2008-04-09 12 -172 val_172 172 val_172 2008-04-08 11 -172 val_172 172 val_172 2008-04-08 12 172 val_172 172 val_172 2008-04-09 12 -172 val_172 172 val_172 2008-04-09 11 -174 val_174 174 val_174 2008-04-08 12 -174 val_174 174 val_174 2008-04-09 12 -174 val_174 174 val_174 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 12 +172 val_172 172 val_172 2008-04-09 12 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 11 -174 val_174 174 val_174 2008-04-08 12 -174 val_174 174 val_174 2008-04-09 11 -174 val_174 174 val_174 2008-04-09 11 -174 val_174 174 val_174 2008-04-08 12 -174 val_174 174 val_174 2008-04-09 12 -174 val_174 174 val_174 2008-04-09 12 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 11 174 val_174 174 val_174 2008-04-08 12 +174 val_174 174 val_174 2008-04-08 12 +174 val_174 174 val_174 2008-04-08 12 +174 val_174 174 val_174 2008-04-08 12 174 val_174 174 val_174 2008-04-09 11 174 val_174 174 val_174 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 11 +174 val_174 174 val_174 2008-04-09 11 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 +174 val_174 174 val_174 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-08 12 -175 val_175 175 val_175 2008-04-08 12 -175 val_175 175 val_175 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 12 175 val_175 175 val_175 2008-04-08 11 -175 val_175 175 val_175 2008-04-09 11 -175 val_175 175 val_175 2008-04-09 11 175 val_175 175 val_175 2008-04-08 12 175 val_175 175 val_175 2008-04-08 12 +175 val_175 175 val_175 2008-04-08 12 +175 val_175 175 val_175 2008-04-08 12 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 11 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 +175 val_175 175 val_175 2008-04-09 12 176 val_176 176 val_176 2008-04-08 11 -176 val_176 176 val_176 2008-04-09 12 -176 val_176 176 val_176 2008-04-08 12 -176 val_176 176 val_176 2008-04-09 12 176 val_176 176 val_176 2008-04-08 11 -176 val_176 176 val_176 2008-04-09 11 -176 val_176 176 val_176 2008-04-09 11 -176 val_176 176 val_176 2008-04-08 12 176 val_176 176 val_176 2008-04-08 11 -176 val_176 176 val_176 2008-04-09 12 -176 val_176 176 val_176 2008-04-08 12 -176 val_176 176 val_176 2008-04-09 12 176 val_176 176 val_176 2008-04-08 11 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 +176 val_176 176 val_176 2008-04-08 12 176 val_176 176 val_176 2008-04-09 11 176 val_176 176 val_176 2008-04-09 11 -176 val_176 176 val_176 2008-04-08 12 -177 val_177 177 val_177 2008-04-08 12 +176 val_176 176 val_176 2008-04-09 11 +176 val_176 176 val_176 2008-04-09 11 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 +176 val_176 176 val_176 2008-04-09 12 177 val_177 177 val_177 2008-04-08 11 +177 val_177 177 val_177 2008-04-08 12 177 val_177 177 val_177 2008-04-09 11 177 val_177 177 val_177 2008-04-09 12 +178 val_178 178 val_178 2008-04-08 11 178 val_178 178 val_178 2008-04-08 12 178 val_178 178 val_178 2008-04-09 11 -178 val_178 178 val_178 2008-04-08 11 178 val_178 178 val_178 2008-04-09 12 -179 val_179 179 val_179 2008-04-08 12 -179 val_179 179 val_179 2008-04-09 12 -179 val_179 179 val_179 2008-04-08 12 -179 val_179 179 val_179 2008-04-09 11 -179 val_179 179 val_179 2008-04-09 12 -179 val_179 179 val_179 2008-04-09 11 +179 val_179 179 val_179 2008-04-08 11 +179 val_179 179 val_179 2008-04-08 11 179 val_179 179 val_179 2008-04-08 11 179 val_179 179 val_179 2008-04-08 11 179 val_179 179 val_179 2008-04-08 12 -179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-08 12 +179 val_179 179 val_179 2008-04-08 12 179 val_179 179 val_179 2008-04-08 12 179 val_179 179 val_179 2008-04-09 11 -179 val_179 179 val_179 2008-04-09 12 179 val_179 179 val_179 2008-04-09 11 -179 val_179 179 val_179 2008-04-08 11 -179 val_179 179 val_179 2008-04-08 11 -180 val_180 180 val_180 2008-04-09 11 -180 val_180 180 val_180 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 11 +179 val_179 179 val_179 2008-04-09 11 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +179 val_179 179 val_179 2008-04-09 12 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 11 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-08 12 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 11 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 +18 val_18 18 val_18 2008-04-09 12 180 val_180 180 val_180 2008-04-08 11 180 val_180 180 val_180 2008-04-08 12 -181 val_181 181 val_181 2008-04-09 11 -181 val_181 181 val_181 2008-04-09 12 +180 val_180 180 val_180 2008-04-09 11 +180 val_180 180 val_180 2008-04-09 12 181 val_181 181 val_181 2008-04-08 11 181 val_181 181 val_181 2008-04-08 12 -183 val_183 183 val_183 2008-04-08 12 +181 val_181 181 val_181 2008-04-09 11 +181 val_181 181 val_181 2008-04-09 12 183 val_183 183 val_183 2008-04-08 11 -183 val_183 183 val_183 2008-04-09 12 +183 val_183 183 val_183 2008-04-08 12 183 val_183 183 val_183 2008-04-09 11 +183 val_183 183 val_183 2008-04-09 12 186 val_186 186 val_186 2008-04-08 11 186 val_186 186 val_186 2008-04-08 12 186 val_186 186 val_186 2008-04-09 11 186 val_186 186 val_186 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-08 11 +187 val_187 187 val_187 2008-04-08 11 +187 val_187 187 val_187 2008-04-08 11 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-08 12 187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-09 11 187 val_187 187 val_187 2008-04-08 12 -187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-08 12 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 11 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 +187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-08 11 187 val_187 187 val_187 2008-04-09 12 187 val_187 187 val_187 2008-04-09 12 -187 val_187 187 val_187 2008-04-08 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-09 11 -187 val_187 187 val_187 2008-04-08 12 -189 val_189 189 val_189 2008-04-09 11 -189 val_189 189 val_189 2008-04-09 12 189 val_189 189 val_189 2008-04-08 11 189 val_189 189 val_189 2008-04-08 12 -190 val_190 190 val_190 2008-04-09 12 +189 val_189 189 val_189 2008-04-09 11 +189 val_189 189 val_189 2008-04-09 12 +19 val_19 19 val_19 2008-04-08 11 +19 val_19 19 val_19 2008-04-08 12 +19 val_19 19 val_19 2008-04-09 11 +19 val_19 19 val_19 2008-04-09 12 +190 val_190 190 val_190 2008-04-08 11 190 val_190 190 val_190 2008-04-08 12 190 val_190 190 val_190 2008-04-09 11 -190 val_190 190 val_190 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-09 12 +190 val_190 190 val_190 2008-04-09 12 191 val_191 191 val_191 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-08 12 191 val_191 191 val_191 2008-04-08 11 -191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-08 11 +191 val_191 191 val_191 2008-04-08 11 +191 val_191 191 val_191 2008-04-08 12 +191 val_191 191 val_191 2008-04-08 12 +191 val_191 191 val_191 2008-04-08 12 191 val_191 191 val_191 2008-04-08 12 191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-09 12 -191 val_191 191 val_191 2008-04-08 11 191 val_191 191 val_191 2008-04-09 11 -191 val_191 191 val_191 2008-04-08 12 -191 val_191 191 val_191 2008-04-08 11 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 11 +191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-09 12 +191 val_191 191 val_191 2008-04-09 12 191 val_191 191 val_191 2008-04-09 12 -191 val_191 191 val_191 2008-04-08 12 -192 val_192 192 val_192 2008-04-09 12 -192 val_192 192 val_192 2008-04-08 12 192 val_192 192 val_192 2008-04-08 11 +192 val_192 192 val_192 2008-04-08 12 192 val_192 192 val_192 2008-04-09 11 +192 val_192 192 val_192 2008-04-09 12 +193 val_193 193 val_193 2008-04-08 11 +193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 11 -193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-08 11 -193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-09 12 193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-08 12 +193 val_193 193 val_193 2008-04-09 11 +193 val_193 193 val_193 2008-04-09 11 +193 val_193 193 val_193 2008-04-09 11 +193 val_193 193 val_193 2008-04-09 11 +193 val_193 193 val_193 2008-04-09 11 +193 val_193 193 val_193 2008-04-09 11 193 val_193 193 val_193 2008-04-09 11 193 val_193 193 val_193 2008-04-09 11 193 val_193 193 val_193 2008-04-09 11 193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 11 193 val_193 193 val_193 2008-04-09 12 -193 val_193 193 val_193 2008-04-08 12 -193 val_193 193 val_193 2008-04-08 11 -194 val_194 194 val_194 2008-04-08 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 +193 val_193 193 val_193 2008-04-09 12 194 val_194 194 val_194 2008-04-08 11 -194 val_194 194 val_194 2008-04-09 12 +194 val_194 194 val_194 2008-04-08 12 194 val_194 194 val_194 2008-04-09 11 +194 val_194 194 val_194 2008-04-09 12 195 val_195 195 val_195 2008-04-08 11 -195 val_195 195 val_195 2008-04-09 11 -195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-08 12 -195 val_195 195 val_195 2008-04-09 12 195 val_195 195 val_195 2008-04-08 11 +195 val_195 195 val_195 2008-04-08 11 +195 val_195 195 val_195 2008-04-08 11 +195 val_195 195 val_195 2008-04-08 12 +195 val_195 195 val_195 2008-04-08 12 +195 val_195 195 val_195 2008-04-08 12 195 val_195 195 val_195 2008-04-08 12 195 val_195 195 val_195 2008-04-09 11 -195 val_195 195 val_195 2008-04-08 11 +195 val_195 195 val_195 2008-04-09 11 +195 val_195 195 val_195 2008-04-09 11 195 val_195 195 val_195 2008-04-09 11 195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-08 12 195 val_195 195 val_195 2008-04-09 12 -195 val_195 195 val_195 2008-04-08 11 -195 val_195 195 val_195 2008-04-08 12 -195 val_195 195 val_195 2008-04-09 11 +195 val_195 195 val_195 2008-04-09 12 +195 val_195 195 val_195 2008-04-09 12 +196 val_196 196 val_196 2008-04-08 11 196 val_196 196 val_196 2008-04-08 12 196 val_196 196 val_196 2008-04-09 11 -196 val_196 196 val_196 2008-04-08 11 196 val_196 196 val_196 2008-04-09 12 -197 val_197 197 val_197 2008-04-08 12 197 val_197 197 val_197 2008-04-08 11 -197 val_197 197 val_197 2008-04-09 11 -197 val_197 197 val_197 2008-04-09 11 +197 val_197 197 val_197 2008-04-08 11 +197 val_197 197 val_197 2008-04-08 11 197 val_197 197 val_197 2008-04-08 11 197 val_197 197 val_197 2008-04-08 12 -197 val_197 197 val_197 2008-04-09 12 -197 val_197 197 val_197 2008-04-09 12 197 val_197 197 val_197 2008-04-08 12 -197 val_197 197 val_197 2008-04-08 11 +197 val_197 197 val_197 2008-04-08 12 +197 val_197 197 val_197 2008-04-08 12 +197 val_197 197 val_197 2008-04-09 11 +197 val_197 197 val_197 2008-04-09 11 197 val_197 197 val_197 2008-04-09 11 197 val_197 197 val_197 2008-04-09 11 -197 val_197 197 val_197 2008-04-08 11 -197 val_197 197 val_197 2008-04-08 12 197 val_197 197 val_197 2008-04-09 12 197 val_197 197 val_197 2008-04-09 12 -199 val_199 199 val_199 2008-04-08 12 +197 val_197 197 val_197 2008-04-09 12 +197 val_197 197 val_197 2008-04-09 12 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 +199 val_199 199 val_199 2008-04-08 11 199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 11 -199 val_199 199 val_199 2008-04-09 11 -199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 199 val_199 199 val_199 2008-04-08 12 199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-08 12 +199 val_199 199 val_199 2008-04-09 11 +199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-09 12 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 12 -199 val_199 199 val_199 2008-04-08 11 -199 val_199 199 val_199 2008-04-09 12 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 199 val_199 199 val_199 2008-04-09 11 -199 val_199 199 val_199 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 11 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +199 val_199 199 val_199 2008-04-09 12 +2 val_2 2 val_2 2008-04-08 11 +2 val_2 2 val_2 2008-04-08 12 +2 val_2 2 val_2 2008-04-09 11 +2 val_2 2 val_2 2008-04-09 12 +20 val_20 20 val_20 2008-04-08 11 +20 val_20 20 val_20 2008-04-08 12 +20 val_20 20 val_20 2008-04-09 11 +20 val_20 20 val_20 2008-04-09 12 +200 val_200 200 val_200 2008-04-08 11 200 val_200 200 val_200 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 11 -200 val_200 200 val_200 2008-04-08 12 -200 val_200 200 val_200 2008-04-08 12 -200 val_200 200 val_200 2008-04-09 12 200 val_200 200 val_200 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 11 -200 val_200 200 val_200 2008-04-09 12 200 val_200 200 val_200 2008-04-08 11 -200 val_200 200 val_200 2008-04-09 11 200 val_200 200 val_200 2008-04-08 12 200 val_200 200 val_200 2008-04-08 12 -200 val_200 200 val_200 2008-04-09 12 -200 val_200 200 val_200 2008-04-08 11 +200 val_200 200 val_200 2008-04-08 12 +200 val_200 200 val_200 2008-04-08 12 200 val_200 200 val_200 2008-04-09 11 -201 val_201 201 val_201 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 11 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 +200 val_200 200 val_200 2008-04-09 12 201 val_201 201 val_201 2008-04-08 11 -201 val_201 201 val_201 2008-04-09 12 201 val_201 201 val_201 2008-04-08 12 -202 val_202 202 val_202 2008-04-09 12 +201 val_201 201 val_201 2008-04-09 11 +201 val_201 201 val_201 2008-04-09 12 202 val_202 202 val_202 2008-04-08 11 -202 val_202 202 val_202 2008-04-09 11 202 val_202 202 val_202 2008-04-08 12 -203 val_203 203 val_203 2008-04-09 11 +202 val_202 202 val_202 2008-04-09 11 +202 val_202 202 val_202 2008-04-09 12 +203 val_203 203 val_203 2008-04-08 11 203 val_203 203 val_203 2008-04-08 11 -203 val_203 203 val_203 2008-04-08 12 -203 val_203 203 val_203 2008-04-09 11 203 val_203 203 val_203 2008-04-08 11 -203 val_203 203 val_203 2008-04-09 12 -203 val_203 203 val_203 2008-04-09 12 -203 val_203 203 val_203 2008-04-08 12 -203 val_203 203 val_203 2008-04-09 11 203 val_203 203 val_203 2008-04-08 11 203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-09 11 +203 val_203 203 val_203 2008-04-09 11 +203 val_203 203 val_203 2008-04-09 11 203 val_203 203 val_203 2008-04-09 11 -203 val_203 203 val_203 2008-04-08 11 203 val_203 203 val_203 2008-04-09 12 203 val_203 203 val_203 2008-04-09 12 -203 val_203 203 val_203 2008-04-08 12 +203 val_203 203 val_203 2008-04-09 12 +203 val_203 203 val_203 2008-04-09 12 +205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-08 11 -205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 11 -205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 12 -205 val_205 205 val_205 2008-04-09 11 205 val_205 205 val_205 2008-04-08 11 -205 val_205 205 val_205 2008-04-09 12 205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 11 205 val_205 205 val_205 2008-04-08 12 -205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-08 12 +205 val_205 205 val_205 2008-04-08 12 +205 val_205 205 val_205 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 11 205 val_205 205 val_205 2008-04-09 11 -205 val_205 205 val_205 2008-04-08 11 205 val_205 205 val_205 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 -207 val_207 207 val_207 2008-04-08 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 +205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-09 12 +205 val_205 205 val_205 2008-04-09 12 207 val_207 207 val_207 2008-04-08 11 -207 val_207 207 val_207 2008-04-08 12 207 val_207 207 val_207 2008-04-08 11 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 -207 val_207 207 val_207 2008-04-08 12 -207 val_207 207 val_207 2008-04-09 12 -207 val_207 207 val_207 2008-04-09 11 207 val_207 207 val_207 2008-04-08 11 -207 val_207 207 val_207 2008-04-08 12 207 val_207 207 val_207 2008-04-08 11 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-08 12 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 11 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +207 val_207 207 val_207 2008-04-09 12 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 +208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 12 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 12 -208 val_208 208 val_208 2008-04-09 11 208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-08 12 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 11 +208 val_208 208 val_208 2008-04-09 12 +208 val_208 208 val_208 2008-04-09 12 +208 val_208 208 val_208 2008-04-09 12 +208 val_208 208 val_208 2008-04-09 12 +208 val_208 208 val_208 2008-04-09 12 +208 val_208 208 val_208 2008-04-09 12 208 val_208 208 val_208 2008-04-09 12 -208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-09 12 -208 val_208 208 val_208 2008-04-08 11 -208 val_208 208 val_208 2008-04-09 11 -208 val_208 208 val_208 2008-04-08 11 208 val_208 208 val_208 2008-04-09 12 -209 val_209 209 val_209 2008-04-08 12 -209 val_209 209 val_209 2008-04-09 11 -209 val_209 209 val_209 2008-04-09 12 -209 val_209 209 val_209 2008-04-09 11 209 val_209 209 val_209 2008-04-08 11 209 val_209 209 val_209 2008-04-08 11 -209 val_209 209 val_209 2008-04-09 12 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-08 12 +209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-08 12 209 val_209 209 val_209 2008-04-09 11 -209 val_209 209 val_209 2008-04-09 12 209 val_209 209 val_209 2008-04-09 11 -209 val_209 209 val_209 2008-04-08 11 -209 val_209 209 val_209 2008-04-08 11 +209 val_209 209 val_209 2008-04-09 11 +209 val_209 209 val_209 2008-04-09 11 +209 val_209 209 val_209 2008-04-09 12 +209 val_209 209 val_209 2008-04-09 12 +209 val_209 209 val_209 2008-04-09 12 209 val_209 209 val_209 2008-04-09 12 -209 val_209 209 val_209 2008-04-08 12 213 val_213 213 val_213 2008-04-08 11 -213 val_213 213 val_213 2008-04-09 12 -213 val_213 213 val_213 2008-04-09 11 -213 val_213 213 val_213 2008-04-09 11 213 val_213 213 val_213 2008-04-08 11 -213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-08 11 +213 val_213 213 val_213 2008-04-08 11 +213 val_213 213 val_213 2008-04-08 12 +213 val_213 213 val_213 2008-04-08 12 213 val_213 213 val_213 2008-04-08 12 213 val_213 213 val_213 2008-04-08 12 -213 val_213 213 val_213 2008-04-08 11 -213 val_213 213 val_213 2008-04-09 12 213 val_213 213 val_213 2008-04-09 11 213 val_213 213 val_213 2008-04-09 11 -213 val_213 213 val_213 2008-04-08 11 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 11 +213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-09 12 +213 val_213 213 val_213 2008-04-09 12 213 val_213 213 val_213 2008-04-09 12 -213 val_213 213 val_213 2008-04-08 12 -213 val_213 213 val_213 2008-04-08 12 -214 val_214 214 val_214 2008-04-08 12 214 val_214 214 val_214 2008-04-08 11 +214 val_214 214 val_214 2008-04-08 12 214 val_214 214 val_214 2008-04-09 11 214 val_214 214 val_214 2008-04-09 12 -216 val_216 216 val_216 2008-04-09 11 -216 val_216 216 val_216 2008-04-08 12 216 val_216 216 val_216 2008-04-08 11 -216 val_216 216 val_216 2008-04-09 12 -216 val_216 216 val_216 2008-04-08 12 -216 val_216 216 val_216 2008-04-09 11 -216 val_216 216 val_216 2008-04-09 12 216 val_216 216 val_216 2008-04-08 11 -216 val_216 216 val_216 2008-04-09 11 -216 val_216 216 val_216 2008-04-08 12 216 val_216 216 val_216 2008-04-08 11 -216 val_216 216 val_216 2008-04-09 12 +216 val_216 216 val_216 2008-04-08 11 216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-08 12 +216 val_216 216 val_216 2008-04-09 11 +216 val_216 216 val_216 2008-04-09 11 +216 val_216 216 val_216 2008-04-09 11 216 val_216 216 val_216 2008-04-09 11 216 val_216 216 val_216 2008-04-09 12 -216 val_216 216 val_216 2008-04-08 11 -217 val_217 217 val_217 2008-04-08 12 +216 val_216 216 val_216 2008-04-09 12 +216 val_216 216 val_216 2008-04-09 12 +216 val_216 216 val_216 2008-04-09 12 217 val_217 217 val_217 2008-04-08 11 -217 val_217 217 val_217 2008-04-09 11 -217 val_217 217 val_217 2008-04-08 12 217 val_217 217 val_217 2008-04-08 11 -217 val_217 217 val_217 2008-04-09 11 -217 val_217 217 val_217 2008-04-09 12 -217 val_217 217 val_217 2008-04-09 12 -217 val_217 217 val_217 2008-04-08 12 217 val_217 217 val_217 2008-04-08 11 -217 val_217 217 val_217 2008-04-09 11 -217 val_217 217 val_217 2008-04-08 12 217 val_217 217 val_217 2008-04-08 11 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-08 12 +217 val_217 217 val_217 2008-04-09 11 +217 val_217 217 val_217 2008-04-09 11 +217 val_217 217 val_217 2008-04-09 11 217 val_217 217 val_217 2008-04-09 11 217 val_217 217 val_217 2008-04-09 12 217 val_217 217 val_217 2008-04-09 12 -218 val_218 218 val_218 2008-04-08 12 +217 val_217 217 val_217 2008-04-09 12 +217 val_217 217 val_217 2008-04-09 12 218 val_218 218 val_218 2008-04-08 11 -218 val_218 218 val_218 2008-04-09 12 +218 val_218 218 val_218 2008-04-08 12 218 val_218 218 val_218 2008-04-09 11 +218 val_218 218 val_218 2008-04-09 12 +219 val_219 219 val_219 2008-04-08 11 219 val_219 219 val_219 2008-04-08 11 -219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-09 11 -219 val_219 219 val_219 2008-04-08 12 219 val_219 219 val_219 2008-04-08 11 +219 val_219 219 val_219 2008-04-08 11 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-08 12 +219 val_219 219 val_219 2008-04-09 11 +219 val_219 219 val_219 2008-04-09 11 +219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-08 12 -219 val_219 219 val_219 2008-04-08 11 219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-09 11 -219 val_219 219 val_219 2008-04-08 12 -219 val_219 219 val_219 2008-04-08 11 -219 val_219 219 val_219 2008-04-09 11 219 val_219 219 val_219 2008-04-09 12 -219 val_219 219 val_219 2008-04-08 12 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-08 12 +219 val_219 219 val_219 2008-04-09 12 221 val_221 221 val_221 2008-04-08 11 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-09 11 -221 val_221 221 val_221 2008-04-08 12 -221 val_221 221 val_221 2008-04-09 11 221 val_221 221 val_221 2008-04-08 11 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-08 11 -221 val_221 221 val_221 2008-04-09 12 -221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-08 11 +221 val_221 221 val_221 2008-04-08 12 +221 val_221 221 val_221 2008-04-08 12 +221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-08 12 221 val_221 221 val_221 2008-04-09 11 -221 val_221 221 val_221 2008-04-08 11 +221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-09 11 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +221 val_221 221 val_221 2008-04-09 12 +222 val_222 222 val_222 2008-04-08 11 222 val_222 222 val_222 2008-04-08 12 -222 val_222 222 val_222 2008-04-09 12 222 val_222 222 val_222 2008-04-09 11 -222 val_222 222 val_222 2008-04-08 11 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-09 11 -223 val_223 223 val_223 2008-04-08 12 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-08 12 +222 val_222 222 val_222 2008-04-09 12 +223 val_223 223 val_223 2008-04-08 11 +223 val_223 223 val_223 2008-04-08 11 223 val_223 223 val_223 2008-04-08 11 -223 val_223 223 val_223 2008-04-09 11 223 val_223 223 val_223 2008-04-08 11 -223 val_223 223 val_223 2008-04-09 12 -223 val_223 223 val_223 2008-04-09 11 223 val_223 223 val_223 2008-04-08 12 -223 val_223 223 val_223 2008-04-09 12 223 val_223 223 val_223 2008-04-08 12 -223 val_223 223 val_223 2008-04-08 11 +223 val_223 223 val_223 2008-04-08 12 +223 val_223 223 val_223 2008-04-08 12 223 val_223 223 val_223 2008-04-09 11 -223 val_223 223 val_223 2008-04-08 11 -224 val_224 224 val_224 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 11 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +223 val_223 223 val_223 2008-04-09 12 +224 val_224 224 val_224 2008-04-08 11 224 val_224 224 val_224 2008-04-08 11 -224 val_224 224 val_224 2008-04-08 12 -224 val_224 224 val_224 2008-04-09 12 -224 val_224 224 val_224 2008-04-09 12 -224 val_224 224 val_224 2008-04-09 11 -224 val_224 224 val_224 2008-04-08 12 224 val_224 224 val_224 2008-04-08 11 -224 val_224 224 val_224 2008-04-09 11 224 val_224 224 val_224 2008-04-08 11 224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-08 12 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 11 +224 val_224 224 val_224 2008-04-09 12 +224 val_224 224 val_224 2008-04-09 12 224 val_224 224 val_224 2008-04-09 12 224 val_224 224 val_224 2008-04-09 12 -224 val_224 224 val_224 2008-04-09 11 -224 val_224 224 val_224 2008-04-08 12 -224 val_224 224 val_224 2008-04-08 11 -226 val_226 226 val_226 2008-04-09 12 226 val_226 226 val_226 2008-04-08 11 -226 val_226 226 val_226 2008-04-09 11 226 val_226 226 val_226 2008-04-08 12 -228 val_228 228 val_228 2008-04-09 11 +226 val_226 226 val_226 2008-04-09 11 +226 val_226 226 val_226 2008-04-09 12 228 val_228 228 val_228 2008-04-08 11 -228 val_228 228 val_228 2008-04-09 12 228 val_228 228 val_228 2008-04-08 12 -229 val_229 229 val_229 2008-04-08 12 -229 val_229 229 val_229 2008-04-09 12 +228 val_228 228 val_228 2008-04-09 11 +228 val_228 228 val_228 2008-04-09 12 229 val_229 229 val_229 2008-04-08 11 -229 val_229 229 val_229 2008-04-09 11 -229 val_229 229 val_229 2008-04-08 12 -229 val_229 229 val_229 2008-04-09 11 -229 val_229 229 val_229 2008-04-09 12 229 val_229 229 val_229 2008-04-08 11 -229 val_229 229 val_229 2008-04-08 12 -229 val_229 229 val_229 2008-04-09 12 229 val_229 229 val_229 2008-04-08 11 -229 val_229 229 val_229 2008-04-09 11 +229 val_229 229 val_229 2008-04-08 11 +229 val_229 229 val_229 2008-04-08 12 +229 val_229 229 val_229 2008-04-08 12 229 val_229 229 val_229 2008-04-08 12 +229 val_229 229 val_229 2008-04-08 12 +229 val_229 229 val_229 2008-04-09 11 +229 val_229 229 val_229 2008-04-09 11 +229 val_229 229 val_229 2008-04-09 11 229 val_229 229 val_229 2008-04-09 11 229 val_229 229 val_229 2008-04-09 12 -229 val_229 229 val_229 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 +229 val_229 229 val_229 2008-04-09 12 +229 val_229 229 val_229 2008-04-09 12 +229 val_229 229 val_229 2008-04-09 12 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 11 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 11 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-08 11 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-08 12 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-09 11 230 val_230 230 val_230 2008-04-09 12 230 val_230 230 val_230 2008-04-09 12 -230 val_230 230 val_230 2008-04-08 11 -230 val_230 230 val_230 2008-04-09 11 -230 val_230 230 val_230 2008-04-08 12 -230 val_230 230 val_230 2008-04-08 11 -233 val_233 233 val_233 2008-04-09 11 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 11 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +230 val_230 230 val_230 2008-04-09 12 +233 val_233 233 val_233 2008-04-08 11 +233 val_233 233 val_233 2008-04-08 11 233 val_233 233 val_233 2008-04-08 11 233 val_233 233 val_233 2008-04-08 11 233 val_233 233 val_233 2008-04-08 12 233 val_233 233 val_233 2008-04-08 12 +233 val_233 233 val_233 2008-04-08 12 +233 val_233 233 val_233 2008-04-08 12 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 11 +233 val_233 233 val_233 2008-04-09 11 233 val_233 233 val_233 2008-04-09 11 233 val_233 233 val_233 2008-04-09 12 233 val_233 233 val_233 2008-04-09 12 -233 val_233 233 val_233 2008-04-09 11 -233 val_233 233 val_233 2008-04-08 11 -233 val_233 233 val_233 2008-04-08 11 -233 val_233 233 val_233 2008-04-08 12 -233 val_233 233 val_233 2008-04-08 12 +233 val_233 233 val_233 2008-04-09 12 +233 val_233 233 val_233 2008-04-09 12 +235 val_235 235 val_235 2008-04-08 11 +235 val_235 235 val_235 2008-04-08 12 235 val_235 235 val_235 2008-04-09 11 235 val_235 235 val_235 2008-04-09 12 -235 val_235 235 val_235 2008-04-08 12 -235 val_235 235 val_235 2008-04-08 11 -237 val_237 237 val_237 2008-04-08 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 11 -237 val_237 237 val_237 2008-04-08 12 237 val_237 237 val_237 2008-04-08 11 237 val_237 237 val_237 2008-04-08 11 -237 val_237 237 val_237 2008-04-09 11 -237 val_237 237 val_237 2008-04-08 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 12 -237 val_237 237 val_237 2008-04-09 11 -237 val_237 237 val_237 2008-04-08 12 237 val_237 237 val_237 2008-04-08 11 237 val_237 237 val_237 2008-04-08 11 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 +237 val_237 237 val_237 2008-04-08 12 237 val_237 237 val_237 2008-04-09 11 -238 val_238 238 val_238 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 11 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 +237 val_237 237 val_237 2008-04-09 12 238 val_238 238 val_238 2008-04-08 11 -238 val_238 238 val_238 2008-04-09 12 238 val_238 238 val_238 2008-04-08 11 -238 val_238 238 val_238 2008-04-08 12 -238 val_238 238 val_238 2008-04-09 11 -238 val_238 238 val_238 2008-04-09 12 -238 val_238 238 val_238 2008-04-08 12 -238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-08 11 -238 val_238 238 val_238 2008-04-09 12 238 val_238 238 val_238 2008-04-08 11 238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-08 12 +238 val_238 238 val_238 2008-04-09 11 +238 val_238 238 val_238 2008-04-09 11 +238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-09 11 238 val_238 238 val_238 2008-04-09 12 -238 val_238 238 val_238 2008-04-08 12 -239 val_239 239 val_239 2008-04-08 12 -239 val_239 239 val_239 2008-04-09 11 -239 val_239 239 val_239 2008-04-09 12 +238 val_238 238 val_238 2008-04-09 12 +238 val_238 238 val_238 2008-04-09 12 +238 val_238 238 val_238 2008-04-09 12 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 11 -239 val_239 239 val_239 2008-04-08 12 -239 val_239 239 val_239 2008-04-09 11 -239 val_239 239 val_239 2008-04-09 12 -239 val_239 239 val_239 2008-04-08 12 -239 val_239 239 val_239 2008-04-09 11 -239 val_239 239 val_239 2008-04-09 12 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 11 239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-08 12 +239 val_239 239 val_239 2008-04-09 11 +239 val_239 239 val_239 2008-04-09 11 +239 val_239 239 val_239 2008-04-09 11 239 val_239 239 val_239 2008-04-09 11 239 val_239 239 val_239 2008-04-09 12 -241 val_241 241 val_241 2008-04-08 12 -241 val_241 241 val_241 2008-04-09 12 +239 val_239 239 val_239 2008-04-09 12 +239 val_239 239 val_239 2008-04-09 12 +239 val_239 239 val_239 2008-04-09 12 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 11 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-08 12 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 11 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 +24 val_24 24 val_24 2008-04-09 12 241 val_241 241 val_241 2008-04-08 11 +241 val_241 241 val_241 2008-04-08 12 241 val_241 241 val_241 2008-04-09 11 +241 val_241 241 val_241 2008-04-09 12 242 val_242 242 val_242 2008-04-08 11 -242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-09 12 -242 val_242 242 val_242 2008-04-08 12 -242 val_242 242 val_242 2008-04-08 12 -242 val_242 242 val_242 2008-04-09 11 242 val_242 242 val_242 2008-04-08 11 -242 val_242 242 val_242 2008-04-09 12 242 val_242 242 val_242 2008-04-08 11 -242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-09 12 +242 val_242 242 val_242 2008-04-08 11 +242 val_242 242 val_242 2008-04-08 12 +242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-08 12 242 val_242 242 val_242 2008-04-09 11 -242 val_242 242 val_242 2008-04-08 11 +242 val_242 242 val_242 2008-04-09 11 +242 val_242 242 val_242 2008-04-09 11 +242 val_242 242 val_242 2008-04-09 11 +242 val_242 242 val_242 2008-04-09 12 +242 val_242 242 val_242 2008-04-09 12 +242 val_242 242 val_242 2008-04-09 12 242 val_242 242 val_242 2008-04-09 12 -244 val_244 244 val_244 2008-04-09 11 244 val_244 244 val_244 2008-04-08 11 244 val_244 244 val_244 2008-04-08 12 +244 val_244 244 val_244 2008-04-09 11 244 val_244 244 val_244 2008-04-09 12 247 val_247 247 val_247 2008-04-08 11 247 val_247 247 val_247 2008-04-08 12 247 val_247 247 val_247 2008-04-09 11 247 val_247 247 val_247 2008-04-09 12 248 val_248 248 val_248 2008-04-08 11 +248 val_248 248 val_248 2008-04-08 12 248 val_248 248 val_248 2008-04-09 11 248 val_248 248 val_248 2008-04-09 12 -248 val_248 248 val_248 2008-04-08 12 -249 val_249 249 val_249 2008-04-09 12 -249 val_249 249 val_249 2008-04-09 11 249 val_249 249 val_249 2008-04-08 11 249 val_249 249 val_249 2008-04-08 12 -252 val_252 252 val_252 2008-04-09 12 -252 val_252 252 val_252 2008-04-09 11 +249 val_249 249 val_249 2008-04-09 11 +249 val_249 249 val_249 2008-04-09 12 252 val_252 252 val_252 2008-04-08 11 252 val_252 252 val_252 2008-04-08 12 +252 val_252 252 val_252 2008-04-09 11 +252 val_252 252 val_252 2008-04-09 12 255 val_255 255 val_255 2008-04-08 11 255 val_255 255 val_255 2008-04-08 11 -255 val_255 255 val_255 2008-04-09 12 -255 val_255 255 val_255 2008-04-08 12 -255 val_255 255 val_255 2008-04-08 12 -255 val_255 255 val_255 2008-04-09 11 -255 val_255 255 val_255 2008-04-09 12 -255 val_255 255 val_255 2008-04-09 11 255 val_255 255 val_255 2008-04-08 11 255 val_255 255 val_255 2008-04-08 11 -255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-08 12 +255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-08 12 255 val_255 255 val_255 2008-04-09 11 -255 val_255 255 val_255 2008-04-09 12 255 val_255 255 val_255 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 11 +255 val_255 255 val_255 2008-04-09 11 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 +255 val_255 255 val_255 2008-04-09 12 256 val_256 256 val_256 2008-04-08 11 256 val_256 256 val_256 2008-04-08 11 -256 val_256 256 val_256 2008-04-08 12 -256 val_256 256 val_256 2008-04-08 12 -256 val_256 256 val_256 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 12 -256 val_256 256 val_256 2008-04-09 11 -256 val_256 256 val_256 2008-04-09 12 256 val_256 256 val_256 2008-04-08 11 256 val_256 256 val_256 2008-04-08 11 256 val_256 256 val_256 2008-04-08 12 256 val_256 256 val_256 2008-04-08 12 +256 val_256 256 val_256 2008-04-08 12 +256 val_256 256 val_256 2008-04-08 12 +256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 11 256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 11 +256 val_256 256 val_256 2008-04-09 12 +256 val_256 256 val_256 2008-04-09 12 +256 val_256 256 val_256 2008-04-09 12 256 val_256 256 val_256 2008-04-09 12 +257 val_257 257 val_257 2008-04-08 11 +257 val_257 257 val_257 2008-04-08 12 257 val_257 257 val_257 2008-04-09 11 257 val_257 257 val_257 2008-04-09 12 -257 val_257 257 val_257 2008-04-08 12 -257 val_257 257 val_257 2008-04-08 11 +258 val_258 258 val_258 2008-04-08 11 258 val_258 258 val_258 2008-04-08 12 258 val_258 258 val_258 2008-04-09 11 258 val_258 258 val_258 2008-04-09 12 -258 val_258 258 val_258 2008-04-08 11 -260 val_260 260 val_260 2008-04-09 12 -260 val_260 260 val_260 2008-04-09 11 -260 val_260 260 val_260 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 11 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-08 12 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 11 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 +26 val_26 26 val_26 2008-04-09 12 260 val_260 260 val_260 2008-04-08 11 +260 val_260 260 val_260 2008-04-08 12 +260 val_260 260 val_260 2008-04-09 11 +260 val_260 260 val_260 2008-04-09 12 262 val_262 262 val_262 2008-04-08 11 -262 val_262 262 val_262 2008-04-09 12 -262 val_262 262 val_262 2008-04-09 11 262 val_262 262 val_262 2008-04-08 12 -263 val_263 263 val_263 2008-04-09 11 +262 val_262 262 val_262 2008-04-09 11 +262 val_262 262 val_262 2008-04-09 12 263 val_263 263 val_263 2008-04-08 11 263 val_263 263 val_263 2008-04-08 12 +263 val_263 263 val_263 2008-04-09 11 263 val_263 263 val_263 2008-04-09 12 265 val_265 265 val_265 2008-04-08 11 -265 val_265 265 val_265 2008-04-08 12 -265 val_265 265 val_265 2008-04-09 11 -265 val_265 265 val_265 2008-04-08 12 -265 val_265 265 val_265 2008-04-09 11 -265 val_265 265 val_265 2008-04-09 12 265 val_265 265 val_265 2008-04-08 11 -265 val_265 265 val_265 2008-04-09 12 +265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-08 12 -265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-08 12 +265 val_265 265 val_265 2008-04-08 12 265 val_265 265 val_265 2008-04-08 12 265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-09 11 +265 val_265 265 val_265 2008-04-09 11 265 val_265 265 val_265 2008-04-09 12 -265 val_265 265 val_265 2008-04-08 11 265 val_265 265 val_265 2008-04-09 12 +265 val_265 265 val_265 2008-04-09 12 +265 val_265 265 val_265 2008-04-09 12 +266 val_266 266 val_266 2008-04-08 11 266 val_266 266 val_266 2008-04-08 12 266 val_266 266 val_266 2008-04-09 11 -266 val_266 266 val_266 2008-04-08 11 266 val_266 266 val_266 2008-04-09 12 -272 val_272 272 val_272 2008-04-09 11 +27 val_27 27 val_27 2008-04-08 11 +27 val_27 27 val_27 2008-04-08 12 +27 val_27 27 val_27 2008-04-09 11 +27 val_27 27 val_27 2008-04-09 12 +272 val_272 272 val_272 2008-04-08 11 272 val_272 272 val_272 2008-04-08 11 -272 val_272 272 val_272 2008-04-09 11 -272 val_272 272 val_272 2008-04-09 12 -272 val_272 272 val_272 2008-04-08 12 -272 val_272 272 val_272 2008-04-09 12 272 val_272 272 val_272 2008-04-08 11 +272 val_272 272 val_272 2008-04-08 11 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-08 12 +272 val_272 272 val_272 2008-04-08 12 272 val_272 272 val_272 2008-04-08 12 272 val_272 272 val_272 2008-04-09 11 -272 val_272 272 val_272 2008-04-08 11 +272 val_272 272 val_272 2008-04-09 11 +272 val_272 272 val_272 2008-04-09 11 272 val_272 272 val_272 2008-04-09 11 272 val_272 272 val_272 2008-04-09 12 -272 val_272 272 val_272 2008-04-08 12 272 val_272 272 val_272 2008-04-09 12 -272 val_272 272 val_272 2008-04-08 11 -272 val_272 272 val_272 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-09 12 +272 val_272 272 val_272 2008-04-09 12 +272 val_272 272 val_272 2008-04-09 12 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 +273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-08 12 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 +273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-08 12 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 -273 val_273 273 val_273 2008-04-08 12 -273 val_273 273 val_273 2008-04-09 11 -273 val_273 273 val_273 2008-04-08 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-09 11 273 val_273 273 val_273 2008-04-09 12 -273 val_273 273 val_273 2008-04-08 12 -274 val_274 274 val_274 2008-04-09 11 -274 val_274 274 val_274 2008-04-09 12 +273 val_273 273 val_273 2008-04-09 12 274 val_274 274 val_274 2008-04-08 11 274 val_274 274 val_274 2008-04-08 12 -275 val_275 275 val_275 2008-04-09 12 +274 val_274 274 val_274 2008-04-09 11 +274 val_274 274 val_274 2008-04-09 12 275 val_275 275 val_275 2008-04-08 11 275 val_275 275 val_275 2008-04-08 12 275 val_275 275 val_275 2008-04-09 11 +275 val_275 275 val_275 2008-04-09 12 +277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 12 -277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-09 12 277 val_277 277 val_277 2008-04-08 11 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-08 11 277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-08 12 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 +277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 11 -277 val_277 277 val_277 2008-04-08 12 277 val_277 277 val_277 2008-04-09 11 277 val_277 277 val_277 2008-04-09 12 -277 val_277 277 val_277 2008-04-08 11 -278 val_278 278 val_278 2008-04-09 11 -278 val_278 278 val_278 2008-04-09 12 -278 val_278 278 val_278 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +277 val_277 277 val_277 2008-04-09 12 +278 val_278 278 val_278 2008-04-08 11 +278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 11 278 val_278 278 val_278 2008-04-08 12 -278 val_278 278 val_278 2008-04-09 11 278 val_278 278 val_278 2008-04-08 12 +278 val_278 278 val_278 2008-04-08 12 +278 val_278 278 val_278 2008-04-08 12 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 11 +278 val_278 278 val_278 2008-04-09 11 278 val_278 278 val_278 2008-04-09 11 278 val_278 278 val_278 2008-04-09 12 278 val_278 278 val_278 2008-04-09 12 -278 val_278 278 val_278 2008-04-08 11 -278 val_278 278 val_278 2008-04-08 11 -278 val_278 278 val_278 2008-04-08 12 -278 val_278 278 val_278 2008-04-09 11 -278 val_278 278 val_278 2008-04-08 12 -280 val_280 280 val_280 2008-04-09 12 -280 val_280 280 val_280 2008-04-08 12 -280 val_280 280 val_280 2008-04-09 11 -280 val_280 280 val_280 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 12 +278 val_278 278 val_278 2008-04-09 12 +28 val_28 28 val_28 2008-04-08 11 +28 val_28 28 val_28 2008-04-08 12 +28 val_28 28 val_28 2008-04-09 11 +28 val_28 28 val_28 2008-04-09 12 280 val_280 280 val_280 2008-04-08 11 -280 val_280 280 val_280 2008-04-08 12 280 val_280 280 val_280 2008-04-08 11 -280 val_280 280 val_280 2008-04-09 11 -280 val_280 280 val_280 2008-04-09 12 -280 val_280 280 val_280 2008-04-08 12 -280 val_280 280 val_280 2008-04-09 11 -280 val_280 280 val_280 2008-04-09 12 280 val_280 280 val_280 2008-04-08 11 -280 val_280 280 val_280 2008-04-08 12 280 val_280 280 val_280 2008-04-08 11 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 +280 val_280 280 val_280 2008-04-08 12 280 val_280 280 val_280 2008-04-09 11 -281 val_281 281 val_281 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 11 +280 val_280 280 val_280 2008-04-09 11 +280 val_280 280 val_280 2008-04-09 11 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 +280 val_280 280 val_280 2008-04-09 12 281 val_281 281 val_281 2008-04-08 11 -281 val_281 281 val_281 2008-04-08 12 -281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-08 11 -281 val_281 281 val_281 2008-04-08 12 -281 val_281 281 val_281 2008-04-09 11 -281 val_281 281 val_281 2008-04-09 12 -281 val_281 281 val_281 2008-04-09 12 281 val_281 281 val_281 2008-04-08 11 -281 val_281 281 val_281 2008-04-08 12 -281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-08 11 281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-08 12 +281 val_281 281 val_281 2008-04-09 11 +281 val_281 281 val_281 2008-04-09 11 281 val_281 281 val_281 2008-04-09 11 +281 val_281 281 val_281 2008-04-09 11 +281 val_281 281 val_281 2008-04-09 12 +281 val_281 281 val_281 2008-04-09 12 +281 val_281 281 val_281 2008-04-09 12 281 val_281 281 val_281 2008-04-09 12 282 val_282 282 val_282 2008-04-08 11 282 val_282 282 val_282 2008-04-08 11 -282 val_282 282 val_282 2008-04-08 12 -282 val_282 282 val_282 2008-04-09 12 -282 val_282 282 val_282 2008-04-09 11 -282 val_282 282 val_282 2008-04-09 11 -282 val_282 282 val_282 2008-04-08 12 -282 val_282 282 val_282 2008-04-09 12 282 val_282 282 val_282 2008-04-08 11 282 val_282 282 val_282 2008-04-08 11 282 val_282 282 val_282 2008-04-08 12 -282 val_282 282 val_282 2008-04-09 12 +282 val_282 282 val_282 2008-04-08 12 +282 val_282 282 val_282 2008-04-08 12 +282 val_282 282 val_282 2008-04-08 12 +282 val_282 282 val_282 2008-04-09 11 +282 val_282 282 val_282 2008-04-09 11 282 val_282 282 val_282 2008-04-09 11 282 val_282 282 val_282 2008-04-09 11 -282 val_282 282 val_282 2008-04-08 12 282 val_282 282 val_282 2008-04-09 12 -283 val_283 283 val_283 2008-04-09 12 +282 val_282 282 val_282 2008-04-09 12 +282 val_282 282 val_282 2008-04-09 12 +282 val_282 282 val_282 2008-04-09 12 283 val_283 283 val_283 2008-04-08 11 283 val_283 283 val_283 2008-04-08 12 283 val_283 283 val_283 2008-04-09 11 +283 val_283 283 val_283 2008-04-09 12 284 val_284 284 val_284 2008-04-08 11 -284 val_284 284 val_284 2008-04-09 12 284 val_284 284 val_284 2008-04-08 12 284 val_284 284 val_284 2008-04-09 11 -285 val_285 285 val_285 2008-04-09 12 -285 val_285 285 val_285 2008-04-09 11 +284 val_284 284 val_284 2008-04-09 12 285 val_285 285 val_285 2008-04-08 11 285 val_285 285 val_285 2008-04-08 12 +285 val_285 285 val_285 2008-04-09 11 +285 val_285 285 val_285 2008-04-09 12 286 val_286 286 val_286 2008-04-08 11 286 val_286 286 val_286 2008-04-08 12 -286 val_286 286 val_286 2008-04-09 12 286 val_286 286 val_286 2008-04-09 11 +286 val_286 286 val_286 2008-04-09 12 287 val_287 287 val_287 2008-04-08 11 -287 val_287 287 val_287 2008-04-09 12 -287 val_287 287 val_287 2008-04-09 11 287 val_287 287 val_287 2008-04-08 12 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 12 -288 val_288 288 val_288 2008-04-08 12 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 12 +287 val_287 287 val_287 2008-04-09 11 +287 val_287 287 val_287 2008-04-09 12 +288 val_288 288 val_288 2008-04-08 11 +288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 11 288 val_288 288 val_288 2008-04-08 12 -288 val_288 288 val_288 2008-04-09 11 -288 val_288 288 val_288 2008-04-09 12 288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 11 +288 val_288 288 val_288 2008-04-09 11 288 val_288 288 val_288 2008-04-09 11 288 val_288 288 val_288 2008-04-09 12 -288 val_288 288 val_288 2008-04-08 11 -288 val_288 288 val_288 2008-04-08 11 -288 val_288 288 val_288 2008-04-08 12 +288 val_288 288 val_288 2008-04-09 12 +288 val_288 288 val_288 2008-04-09 12 +288 val_288 288 val_288 2008-04-09 12 +289 val_289 289 val_289 2008-04-08 11 289 val_289 289 val_289 2008-04-08 12 289 val_289 289 val_289 2008-04-09 11 -289 val_289 289 val_289 2008-04-08 11 289 val_289 289 val_289 2008-04-09 12 -291 val_291 291 val_291 2008-04-09 12 +291 val_291 291 val_291 2008-04-08 11 291 val_291 291 val_291 2008-04-08 12 291 val_291 291 val_291 2008-04-09 11 -291 val_291 291 val_291 2008-04-08 11 -292 val_292 292 val_292 2008-04-09 11 +291 val_291 291 val_291 2008-04-09 12 292 val_292 292 val_292 2008-04-08 11 -292 val_292 292 val_292 2008-04-09 12 292 val_292 292 val_292 2008-04-08 12 -296 val_296 296 val_296 2008-04-08 12 +292 val_292 292 val_292 2008-04-09 11 +292 val_292 292 val_292 2008-04-09 12 296 val_296 296 val_296 2008-04-08 11 -296 val_296 296 val_296 2008-04-09 12 +296 val_296 296 val_296 2008-04-08 12 296 val_296 296 val_296 2008-04-09 11 +296 val_296 296 val_296 2008-04-09 12 298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 11 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 +298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 12 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-09 12 298 val_298 298 val_298 2008-04-09 11 -298 val_298 298 val_298 2008-04-08 11 298 val_298 298 val_298 2008-04-09 12 -298 val_298 298 val_298 2008-04-08 11 -298 val_298 298 val_298 2008-04-08 12 -302 val_302 302 val_302 2008-04-08 12 -302 val_302 302 val_302 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +298 val_298 298 val_298 2008-04-09 12 +30 val_30 30 val_30 2008-04-08 11 +30 val_30 30 val_30 2008-04-08 12 +30 val_30 30 val_30 2008-04-09 11 +30 val_30 30 val_30 2008-04-09 12 302 val_302 302 val_302 2008-04-08 11 +302 val_302 302 val_302 2008-04-08 12 302 val_302 302 val_302 2008-04-09 11 +302 val_302 302 val_302 2008-04-09 12 305 val_305 305 val_305 2008-04-08 11 -305 val_305 305 val_305 2008-04-09 12 -305 val_305 305 val_305 2008-04-09 11 305 val_305 305 val_305 2008-04-08 12 -306 val_306 306 val_306 2008-04-09 11 +305 val_305 305 val_305 2008-04-09 11 +305 val_305 305 val_305 2008-04-09 12 306 val_306 306 val_306 2008-04-08 11 -306 val_306 306 val_306 2008-04-09 12 306 val_306 306 val_306 2008-04-08 12 +306 val_306 306 val_306 2008-04-09 11 +306 val_306 306 val_306 2008-04-09 12 307 val_307 307 val_307 2008-04-08 11 -307 val_307 307 val_307 2008-04-08 12 -307 val_307 307 val_307 2008-04-09 12 -307 val_307 307 val_307 2008-04-09 12 307 val_307 307 val_307 2008-04-08 11 -307 val_307 307 val_307 2008-04-08 12 -307 val_307 307 val_307 2008-04-09 11 -307 val_307 307 val_307 2008-04-09 11 307 val_307 307 val_307 2008-04-08 11 -307 val_307 307 val_307 2008-04-08 12 -307 val_307 307 val_307 2008-04-09 12 -307 val_307 307 val_307 2008-04-09 12 307 val_307 307 val_307 2008-04-08 11 307 val_307 307 val_307 2008-04-08 12 +307 val_307 307 val_307 2008-04-08 12 +307 val_307 307 val_307 2008-04-08 12 +307 val_307 307 val_307 2008-04-08 12 307 val_307 307 val_307 2008-04-09 11 307 val_307 307 val_307 2008-04-09 11 +307 val_307 307 val_307 2008-04-09 11 +307 val_307 307 val_307 2008-04-09 11 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 +307 val_307 307 val_307 2008-04-09 12 308 val_308 308 val_308 2008-04-08 11 -308 val_308 308 val_308 2008-04-09 12 308 val_308 308 val_308 2008-04-08 12 308 val_308 308 val_308 2008-04-09 11 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-09 11 +308 val_308 308 val_308 2008-04-09 12 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-08 12 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 11 -309 val_309 309 val_309 2008-04-08 12 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-09 11 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 12 -309 val_309 309 val_309 2008-04-08 12 309 val_309 309 val_309 2008-04-08 11 -309 val_309 309 val_309 2008-04-09 11 309 val_309 309 val_309 2008-04-08 12 -310 val_310 310 val_310 2008-04-09 11 -310 val_310 310 val_310 2008-04-09 12 +309 val_309 309 val_309 2008-04-08 12 +309 val_309 309 val_309 2008-04-08 12 +309 val_309 309 val_309 2008-04-08 12 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 11 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 +309 val_309 309 val_309 2008-04-09 12 310 val_310 310 val_310 2008-04-08 11 310 val_310 310 val_310 2008-04-08 12 +310 val_310 310 val_310 2008-04-09 11 +310 val_310 310 val_310 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-08 12 -311 val_311 311 val_311 2008-04-08 12 -311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-08 12 -311 val_311 311 val_311 2008-04-08 12 -311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 -311 val_311 311 val_311 2008-04-09 12 311 val_311 311 val_311 2008-04-08 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 -311 val_311 311 val_311 2008-04-09 11 311 val_311 311 val_311 2008-04-08 11 311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 12 311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-08 12 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 11 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 +311 val_311 311 val_311 2008-04-09 12 315 val_315 315 val_315 2008-04-08 11 -315 val_315 315 val_315 2008-04-09 12 -315 val_315 315 val_315 2008-04-09 11 315 val_315 315 val_315 2008-04-08 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 12 +315 val_315 315 val_315 2008-04-09 11 +315 val_315 315 val_315 2008-04-09 12 316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-08 11 316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-08 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-08 11 316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 11 -316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-08 12 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 +316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 11 316 val_316 316 val_316 2008-04-09 12 -316 val_316 316 val_316 2008-04-08 11 -316 val_316 316 val_316 2008-04-08 11 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 +316 val_316 316 val_316 2008-04-09 12 316 val_316 316 val_316 2008-04-09 12 317 val_317 317 val_317 2008-04-08 11 -317 val_317 317 val_317 2008-04-09 11 -317 val_317 317 val_317 2008-04-09 12 -317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-08 11 +317 val_317 317 val_317 2008-04-08 11 +317 val_317 317 val_317 2008-04-08 11 +317 val_317 317 val_317 2008-04-08 12 +317 val_317 317 val_317 2008-04-08 12 317 val_317 317 val_317 2008-04-08 12 317 val_317 317 val_317 2008-04-08 12 -317 val_317 317 val_317 2008-04-08 11 317 val_317 317 val_317 2008-04-09 11 -317 val_317 317 val_317 2008-04-08 11 +317 val_317 317 val_317 2008-04-09 11 +317 val_317 317 val_317 2008-04-09 11 317 val_317 317 val_317 2008-04-09 11 317 val_317 317 val_317 2008-04-09 12 317 val_317 317 val_317 2008-04-09 12 -317 val_317 317 val_317 2008-04-08 12 -317 val_317 317 val_317 2008-04-08 12 -317 val_317 317 val_317 2008-04-08 11 -317 val_317 317 val_317 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 12 +317 val_317 317 val_317 2008-04-09 12 +318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-09 12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 11 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-08 12 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 11 318 val_318 318 val_318 2008-04-09 11 318 val_318 318 val_318 2008-04-09 11 -318 val_318 318 val_318 2008-04-09 12 318 val_318 318 val_318 2008-04-09 11 318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-08 11 -318 val_318 318 val_318 2008-04-08 11 318 val_318 318 val_318 2008-04-09 12 -318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 12 -318 val_318 318 val_318 2008-04-08 11 -321 val_321 321 val_321 2008-04-09 11 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 +318 val_318 318 val_318 2008-04-09 12 321 val_321 321 val_321 2008-04-08 11 321 val_321 321 val_321 2008-04-08 11 -321 val_321 321 val_321 2008-04-08 12 -321 val_321 321 val_321 2008-04-08 12 -321 val_321 321 val_321 2008-04-09 11 -321 val_321 321 val_321 2008-04-09 12 -321 val_321 321 val_321 2008-04-09 12 -321 val_321 321 val_321 2008-04-09 11 321 val_321 321 val_321 2008-04-08 11 321 val_321 321 val_321 2008-04-08 11 321 val_321 321 val_321 2008-04-08 12 321 val_321 321 val_321 2008-04-08 12 +321 val_321 321 val_321 2008-04-08 12 +321 val_321 321 val_321 2008-04-08 12 +321 val_321 321 val_321 2008-04-09 11 +321 val_321 321 val_321 2008-04-09 11 +321 val_321 321 val_321 2008-04-09 11 321 val_321 321 val_321 2008-04-09 11 321 val_321 321 val_321 2008-04-09 12 321 val_321 321 val_321 2008-04-09 12 -322 val_322 322 val_322 2008-04-08 12 +321 val_321 321 val_321 2008-04-09 12 +321 val_321 321 val_321 2008-04-09 12 +322 val_322 322 val_322 2008-04-08 11 322 val_322 322 val_322 2008-04-08 11 -322 val_322 322 val_322 2008-04-09 11 -322 val_322 322 val_322 2008-04-09 12 -322 val_322 322 val_322 2008-04-09 12 -322 val_322 322 val_322 2008-04-09 11 -322 val_322 322 val_322 2008-04-08 12 322 val_322 322 val_322 2008-04-08 11 -322 val_322 322 val_322 2008-04-08 12 322 val_322 322 val_322 2008-04-08 11 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-08 12 +322 val_322 322 val_322 2008-04-09 11 +322 val_322 322 val_322 2008-04-09 11 +322 val_322 322 val_322 2008-04-09 11 322 val_322 322 val_322 2008-04-09 11 322 val_322 322 val_322 2008-04-09 12 322 val_322 322 val_322 2008-04-09 12 -322 val_322 322 val_322 2008-04-09 11 -322 val_322 322 val_322 2008-04-08 12 -322 val_322 322 val_322 2008-04-08 11 +322 val_322 322 val_322 2008-04-09 12 +322 val_322 322 val_322 2008-04-09 12 323 val_323 323 val_323 2008-04-08 11 323 val_323 323 val_323 2008-04-08 12 323 val_323 323 val_323 2008-04-09 11 323 val_323 323 val_323 2008-04-09 12 -325 val_325 325 val_325 2008-04-09 11 -325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-09 12 -325 val_325 325 val_325 2008-04-09 11 -325 val_325 325 val_325 2008-04-09 12 -325 val_325 325 val_325 2008-04-08 12 325 val_325 325 val_325 2008-04-08 11 325 val_325 325 val_325 2008-04-08 11 -325 val_325 325 val_325 2008-04-09 11 +325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-08 11 325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-09 12 +325 val_325 325 val_325 2008-04-08 12 +325 val_325 325 val_325 2008-04-08 12 +325 val_325 325 val_325 2008-04-08 12 +325 val_325 325 val_325 2008-04-09 11 +325 val_325 325 val_325 2008-04-09 11 +325 val_325 325 val_325 2008-04-09 11 325 val_325 325 val_325 2008-04-09 11 325 val_325 325 val_325 2008-04-09 12 -325 val_325 325 val_325 2008-04-08 12 -325 val_325 325 val_325 2008-04-08 11 -325 val_325 325 val_325 2008-04-08 11 +325 val_325 325 val_325 2008-04-09 12 +325 val_325 325 val_325 2008-04-09 12 +325 val_325 325 val_325 2008-04-09 12 327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-09 12 -327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-08 12 -327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-09 12 -327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 11 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-08 12 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 11 -327 val_327 327 val_327 2008-04-08 12 +327 val_327 327 val_327 2008-04-09 11 +327 val_327 327 val_327 2008-04-09 11 +327 val_327 327 val_327 2008-04-09 11 +327 val_327 327 val_327 2008-04-09 11 +327 val_327 327 val_327 2008-04-09 11 +327 val_327 327 val_327 2008-04-09 11 327 val_327 327 val_327 2008-04-09 12 -327 val_327 327 val_327 2008-04-08 11 -327 val_327 327 val_327 2008-04-08 11 327 val_327 327 val_327 2008-04-09 12 327 val_327 327 val_327 2008-04-09 12 -331 val_331 331 val_331 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +327 val_327 327 val_327 2008-04-09 12 +33 val_33 33 val_33 2008-04-08 11 +33 val_33 33 val_33 2008-04-08 12 +33 val_33 33 val_33 2008-04-09 11 +33 val_33 33 val_33 2008-04-09 12 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 11 -331 val_331 331 val_331 2008-04-08 12 -331 val_331 331 val_331 2008-04-09 11 -331 val_331 331 val_331 2008-04-09 11 -331 val_331 331 val_331 2008-04-08 12 -331 val_331 331 val_331 2008-04-09 12 -331 val_331 331 val_331 2008-04-09 12 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 11 331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-08 12 331 val_331 331 val_331 2008-04-09 11 331 val_331 331 val_331 2008-04-09 11 -331 val_331 331 val_331 2008-04-08 12 +331 val_331 331 val_331 2008-04-09 11 +331 val_331 331 val_331 2008-04-09 11 +331 val_331 331 val_331 2008-04-09 12 +331 val_331 331 val_331 2008-04-09 12 +331 val_331 331 val_331 2008-04-09 12 331 val_331 331 val_331 2008-04-09 12 332 val_332 332 val_332 2008-04-08 11 -332 val_332 332 val_332 2008-04-09 11 332 val_332 332 val_332 2008-04-08 12 +332 val_332 332 val_332 2008-04-09 11 332 val_332 332 val_332 2008-04-09 12 -333 val_333 333 val_333 2008-04-09 11 -333 val_333 333 val_333 2008-04-08 12 -333 val_333 333 val_333 2008-04-09 12 -333 val_333 333 val_333 2008-04-08 12 333 val_333 333 val_333 2008-04-08 11 -333 val_333 333 val_333 2008-04-09 11 333 val_333 333 val_333 2008-04-08 11 -333 val_333 333 val_333 2008-04-09 12 -333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-08 12 +333 val_333 333 val_333 2008-04-08 12 333 val_333 333 val_333 2008-04-08 12 -333 val_333 333 val_333 2008-04-09 12 333 val_333 333 val_333 2008-04-08 12 -333 val_333 333 val_333 2008-04-08 11 333 val_333 333 val_333 2008-04-09 11 -333 val_333 333 val_333 2008-04-08 11 +333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-09 11 +333 val_333 333 val_333 2008-04-09 12 333 val_333 333 val_333 2008-04-09 12 +333 val_333 333 val_333 2008-04-09 12 +333 val_333 333 val_333 2008-04-09 12 +335 val_335 335 val_335 2008-04-08 11 335 val_335 335 val_335 2008-04-08 12 -335 val_335 335 val_335 2008-04-09 12 335 val_335 335 val_335 2008-04-09 11 -335 val_335 335 val_335 2008-04-08 11 +335 val_335 335 val_335 2008-04-09 12 +336 val_336 336 val_336 2008-04-08 11 +336 val_336 336 val_336 2008-04-08 12 336 val_336 336 val_336 2008-04-09 11 336 val_336 336 val_336 2008-04-09 12 -336 val_336 336 val_336 2008-04-08 12 -336 val_336 336 val_336 2008-04-08 11 338 val_338 338 val_338 2008-04-08 11 -338 val_338 338 val_338 2008-04-09 11 338 val_338 338 val_338 2008-04-08 12 +338 val_338 338 val_338 2008-04-09 11 338 val_338 338 val_338 2008-04-09 12 -339 val_339 339 val_339 2008-04-09 11 339 val_339 339 val_339 2008-04-08 11 -339 val_339 339 val_339 2008-04-09 12 339 val_339 339 val_339 2008-04-08 12 -341 val_341 341 val_341 2008-04-09 12 -341 val_341 341 val_341 2008-04-09 11 +339 val_339 339 val_339 2008-04-09 11 +339 val_339 339 val_339 2008-04-09 12 +34 val_34 34 val_34 2008-04-08 11 +34 val_34 34 val_34 2008-04-08 12 +34 val_34 34 val_34 2008-04-09 11 +34 val_34 34 val_34 2008-04-09 12 341 val_341 341 val_341 2008-04-08 11 341 val_341 341 val_341 2008-04-08 12 +341 val_341 341 val_341 2008-04-09 11 +341 val_341 341 val_341 2008-04-09 12 +342 val_342 342 val_342 2008-04-08 11 342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 -342 val_342 342 val_342 2008-04-08 12 -342 val_342 342 val_342 2008-04-09 11 -342 val_342 342 val_342 2008-04-08 12 342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 -342 val_342 342 val_342 2008-04-09 11 342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 342 val_342 342 val_342 2008-04-08 12 -342 val_342 342 val_342 2008-04-09 11 342 val_342 342 val_342 2008-04-08 12 -342 val_342 342 val_342 2008-04-08 11 -342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-08 12 +342 val_342 342 val_342 2008-04-08 12 342 val_342 342 val_342 2008-04-09 11 -344 val_344 344 val_344 2008-04-08 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 11 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 +342 val_342 342 val_342 2008-04-09 12 344 val_344 344 val_344 2008-04-08 11 -344 val_344 344 val_344 2008-04-08 12 -344 val_344 344 val_344 2008-04-09 11 344 val_344 344 val_344 2008-04-08 11 -344 val_344 344 val_344 2008-04-08 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 12 -344 val_344 344 val_344 2008-04-09 11 344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-08 12 +344 val_344 344 val_344 2008-04-08 12 +344 val_344 344 val_344 2008-04-08 12 344 val_344 344 val_344 2008-04-08 12 344 val_344 344 val_344 2008-04-09 11 -344 val_344 344 val_344 2008-04-08 11 +344 val_344 344 val_344 2008-04-09 11 +344 val_344 344 val_344 2008-04-09 11 +344 val_344 344 val_344 2008-04-09 11 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 +344 val_344 344 val_344 2008-04-09 12 345 val_345 345 val_345 2008-04-08 11 -345 val_345 345 val_345 2008-04-09 11 345 val_345 345 val_345 2008-04-08 12 +345 val_345 345 val_345 2008-04-09 11 345 val_345 345 val_345 2008-04-09 12 -348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 +348 val_348 348 val_348 2008-04-08 12 348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-09 12 348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 11 -348 val_348 348 val_348 2008-04-08 12 -348 val_348 348 val_348 2008-04-08 11 +348 val_348 348 val_348 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 11 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 12 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 348 val_348 348 val_348 2008-04-09 12 -348 val_348 348 val_348 2008-04-08 11 -351 val_351 351 val_351 2008-04-09 11 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +348 val_348 348 val_348 2008-04-09 12 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 11 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-08 12 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 11 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 +35 val_35 35 val_35 2008-04-09 12 351 val_351 351 val_351 2008-04-08 11 351 val_351 351 val_351 2008-04-08 12 +351 val_351 351 val_351 2008-04-09 11 351 val_351 351 val_351 2008-04-09 12 -353 val_353 353 val_353 2008-04-08 12 -353 val_353 353 val_353 2008-04-08 12 -353 val_353 353 val_353 2008-04-09 11 -353 val_353 353 val_353 2008-04-09 12 353 val_353 353 val_353 2008-04-08 11 -353 val_353 353 val_353 2008-04-09 11 -353 val_353 353 val_353 2008-04-09 12 353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-08 12 +353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-08 12 353 val_353 353 val_353 2008-04-09 11 -353 val_353 353 val_353 2008-04-09 12 -353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-09 11 +353 val_353 353 val_353 2008-04-09 11 353 val_353 353 val_353 2008-04-09 11 353 val_353 353 val_353 2008-04-09 12 -353 val_353 353 val_353 2008-04-08 11 +353 val_353 353 val_353 2008-04-09 12 +353 val_353 353 val_353 2008-04-09 12 +353 val_353 353 val_353 2008-04-09 12 +356 val_356 356 val_356 2008-04-08 11 356 val_356 356 val_356 2008-04-08 12 356 val_356 356 val_356 2008-04-09 11 -356 val_356 356 val_356 2008-04-08 11 356 val_356 356 val_356 2008-04-09 12 360 val_360 360 val_360 2008-04-08 11 360 val_360 360 val_360 2008-04-08 12 -360 val_360 360 val_360 2008-04-09 12 360 val_360 360 val_360 2008-04-09 11 -362 val_362 362 val_362 2008-04-09 12 +360 val_360 360 val_360 2008-04-09 12 362 val_362 362 val_362 2008-04-08 11 -362 val_362 362 val_362 2008-04-09 11 362 val_362 362 val_362 2008-04-08 12 -364 val_364 364 val_364 2008-04-09 12 +362 val_362 362 val_362 2008-04-09 11 +362 val_362 362 val_362 2008-04-09 12 364 val_364 364 val_364 2008-04-08 11 -364 val_364 364 val_364 2008-04-09 11 364 val_364 364 val_364 2008-04-08 12 +364 val_364 364 val_364 2008-04-09 11 +364 val_364 364 val_364 2008-04-09 12 365 val_365 365 val_365 2008-04-08 11 -365 val_365 365 val_365 2008-04-09 12 365 val_365 365 val_365 2008-04-08 12 365 val_365 365 val_365 2008-04-09 11 +365 val_365 365 val_365 2008-04-09 12 +366 val_366 366 val_366 2008-04-08 11 366 val_366 366 val_366 2008-04-08 12 -366 val_366 366 val_366 2008-04-09 12 366 val_366 366 val_366 2008-04-09 11 -366 val_366 366 val_366 2008-04-08 11 -367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 11 +366 val_366 366 val_366 2008-04-09 12 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 11 -367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 11 -367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 -367 val_367 367 val_367 2008-04-09 11 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 11 367 val_367 367 val_367 2008-04-08 12 -367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-08 12 +367 val_367 367 val_367 2008-04-08 12 +367 val_367 367 val_367 2008-04-08 12 367 val_367 367 val_367 2008-04-09 11 -368 val_368 368 val_368 2008-04-08 12 +367 val_367 367 val_367 2008-04-09 11 +367 val_367 367 val_367 2008-04-09 11 +367 val_367 367 val_367 2008-04-09 11 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 +367 val_367 367 val_367 2008-04-09 12 368 val_368 368 val_368 2008-04-08 11 -368 val_368 368 val_368 2008-04-09 12 +368 val_368 368 val_368 2008-04-08 12 368 val_368 368 val_368 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-08 12 +368 val_368 368 val_368 2008-04-09 12 +369 val_369 369 val_369 2008-04-08 11 +369 val_369 369 val_369 2008-04-08 11 +369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-09 11 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-08 12 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-09 11 369 val_369 369 val_369 2008-04-08 11 369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-09 12 369 val_369 369 val_369 2008-04-08 12 -369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-08 12 369 val_369 369 val_369 2008-04-09 11 369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-09 12 -369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 369 val_369 369 val_369 2008-04-09 11 -369 val_369 369 val_369 2008-04-08 11 -369 val_369 369 val_369 2008-04-08 12 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 11 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +369 val_369 369 val_369 2008-04-09 12 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 11 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-08 12 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 11 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 +37 val_37 37 val_37 2008-04-09 12 373 val_373 373 val_373 2008-04-08 11 +373 val_373 373 val_373 2008-04-08 12 373 val_373 373 val_373 2008-04-09 11 373 val_373 373 val_373 2008-04-09 12 -373 val_373 373 val_373 2008-04-08 12 -374 val_374 374 val_374 2008-04-09 12 -374 val_374 374 val_374 2008-04-08 12 374 val_374 374 val_374 2008-04-08 11 +374 val_374 374 val_374 2008-04-08 12 374 val_374 374 val_374 2008-04-09 11 -375 val_375 375 val_375 2008-04-09 12 -375 val_375 375 val_375 2008-04-09 11 -375 val_375 375 val_375 2008-04-08 12 +374 val_374 374 val_374 2008-04-09 12 375 val_375 375 val_375 2008-04-08 11 -377 val_377 377 val_377 2008-04-09 11 -377 val_377 377 val_377 2008-04-09 12 +375 val_375 375 val_375 2008-04-08 12 +375 val_375 375 val_375 2008-04-09 11 +375 val_375 375 val_375 2008-04-09 12 377 val_377 377 val_377 2008-04-08 11 377 val_377 377 val_377 2008-04-08 12 +377 val_377 377 val_377 2008-04-09 11 +377 val_377 377 val_377 2008-04-09 12 +378 val_378 378 val_378 2008-04-08 11 +378 val_378 378 val_378 2008-04-08 12 378 val_378 378 val_378 2008-04-09 11 378 val_378 378 val_378 2008-04-09 12 -378 val_378 378 val_378 2008-04-08 12 -378 val_378 378 val_378 2008-04-08 11 -379 val_379 379 val_379 2008-04-08 12 379 val_379 379 val_379 2008-04-08 11 +379 val_379 379 val_379 2008-04-08 12 379 val_379 379 val_379 2008-04-09 11 379 val_379 379 val_379 2008-04-09 12 -382 val_382 382 val_382 2008-04-08 12 -382 val_382 382 val_382 2008-04-09 11 -382 val_382 382 val_382 2008-04-09 11 -382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-08 11 -382 val_382 382 val_382 2008-04-09 12 -382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-08 12 +382 val_382 382 val_382 2008-04-08 12 +382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-08 12 382 val_382 382 val_382 2008-04-09 11 382 val_382 382 val_382 2008-04-09 11 -382 val_382 382 val_382 2008-04-08 12 -382 val_382 382 val_382 2008-04-08 11 +382 val_382 382 val_382 2008-04-09 11 +382 val_382 382 val_382 2008-04-09 11 +382 val_382 382 val_382 2008-04-09 12 +382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-09 12 382 val_382 382 val_382 2008-04-09 12 -382 val_382 382 val_382 2008-04-08 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-09 12 384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 11 384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 +384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-08 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 11 384 val_384 384 val_384 2008-04-09 12 384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-08 12 384 val_384 384 val_384 2008-04-09 12 -384 val_384 384 val_384 2008-04-08 12 -384 val_384 384 val_384 2008-04-09 11 -384 val_384 384 val_384 2008-04-08 11 -384 val_384 384 val_384 2008-04-09 11 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +384 val_384 384 val_384 2008-04-09 12 +386 val_386 386 val_386 2008-04-08 11 386 val_386 386 val_386 2008-04-08 12 -386 val_386 386 val_386 2008-04-09 12 386 val_386 386 val_386 2008-04-09 11 -386 val_386 386 val_386 2008-04-08 11 -389 val_389 389 val_389 2008-04-08 12 -389 val_389 389 val_389 2008-04-09 12 +386 val_386 386 val_386 2008-04-09 12 389 val_389 389 val_389 2008-04-08 11 +389 val_389 389 val_389 2008-04-08 12 389 val_389 389 val_389 2008-04-09 11 -392 val_392 392 val_392 2008-04-08 12 +389 val_389 389 val_389 2008-04-09 12 392 val_392 392 val_392 2008-04-08 11 -392 val_392 392 val_392 2008-04-09 12 +392 val_392 392 val_392 2008-04-08 12 392 val_392 392 val_392 2008-04-09 11 -393 val_393 393 val_393 2008-04-09 12 -393 val_393 393 val_393 2008-04-08 12 +392 val_392 392 val_392 2008-04-09 12 393 val_393 393 val_393 2008-04-08 11 +393 val_393 393 val_393 2008-04-08 12 393 val_393 393 val_393 2008-04-09 11 -394 val_394 394 val_394 2008-04-08 12 -394 val_394 394 val_394 2008-04-09 12 +393 val_393 393 val_393 2008-04-09 12 394 val_394 394 val_394 2008-04-08 11 +394 val_394 394 val_394 2008-04-08 12 394 val_394 394 val_394 2008-04-09 11 -395 val_395 395 val_395 2008-04-09 12 -395 val_395 395 val_395 2008-04-09 11 -395 val_395 395 val_395 2008-04-09 11 -395 val_395 395 val_395 2008-04-08 12 +394 val_394 394 val_394 2008-04-09 12 395 val_395 395 val_395 2008-04-08 11 -395 val_395 395 val_395 2008-04-09 12 -395 val_395 395 val_395 2008-04-08 12 395 val_395 395 val_395 2008-04-08 11 -395 val_395 395 val_395 2008-04-09 12 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-08 12 +395 val_395 395 val_395 2008-04-09 11 +395 val_395 395 val_395 2008-04-09 11 395 val_395 395 val_395 2008-04-09 11 395 val_395 395 val_395 2008-04-09 11 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 11 395 val_395 395 val_395 2008-04-09 12 -395 val_395 395 val_395 2008-04-08 12 -395 val_395 395 val_395 2008-04-08 11 +395 val_395 395 val_395 2008-04-09 12 +395 val_395 395 val_395 2008-04-09 12 +395 val_395 395 val_395 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 -396 val_396 396 val_396 2008-04-09 12 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-09 11 -396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 11 396 val_396 396 val_396 2008-04-08 12 -396 val_396 396 val_396 2008-04-09 11 396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-08 12 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 11 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 +396 val_396 396 val_396 2008-04-09 12 396 val_396 396 val_396 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 11 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 11 397 val_397 397 val_397 2008-04-08 11 397 val_397 397 val_397 2008-04-08 11 +397 val_397 397 val_397 2008-04-08 11 +397 val_397 397 val_397 2008-04-08 11 +397 val_397 397 val_397 2008-04-08 12 +397 val_397 397 val_397 2008-04-08 12 397 val_397 397 val_397 2008-04-08 12 397 val_397 397 val_397 2008-04-08 12 397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 11 +397 val_397 397 val_397 2008-04-09 12 +397 val_397 397 val_397 2008-04-09 12 397 val_397 397 val_397 2008-04-09 12 397 val_397 397 val_397 2008-04-09 12 -397 val_397 397 val_397 2008-04-09 11 -397 val_397 397 val_397 2008-04-08 11 -397 val_397 397 val_397 2008-04-08 11 -397 val_397 397 val_397 2008-04-08 12 -397 val_397 397 val_397 2008-04-08 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 11 -399 val_399 399 val_399 2008-04-08 12 -399 val_399 399 val_399 2008-04-09 11 399 val_399 399 val_399 2008-04-08 11 -399 val_399 399 val_399 2008-04-08 12 399 val_399 399 val_399 2008-04-08 11 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 12 -399 val_399 399 val_399 2008-04-09 11 -399 val_399 399 val_399 2008-04-08 12 -399 val_399 399 val_399 2008-04-09 11 399 val_399 399 val_399 2008-04-08 11 -399 val_399 399 val_399 2008-04-08 12 399 val_399 399 val_399 2008-04-08 11 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-08 12 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 11 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +399 val_399 399 val_399 2008-04-09 12 +4 val_4 4 val_4 2008-04-08 11 +4 val_4 4 val_4 2008-04-08 12 +4 val_4 4 val_4 2008-04-09 11 +4 val_4 4 val_4 2008-04-09 12 +400 val_400 400 val_400 2008-04-08 11 400 val_400 400 val_400 2008-04-08 12 400 val_400 400 val_400 2008-04-09 11 400 val_400 400 val_400 2008-04-09 12 -400 val_400 400 val_400 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 +401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-08 12 -401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 12 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 11 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 12 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 11 401 val_401 401 val_401 2008-04-09 12 -401 val_401 401 val_401 2008-04-08 11 -401 val_401 401 val_401 2008-04-08 12 401 val_401 401 val_401 2008-04-09 12 -402 val_402 402 val_402 2008-04-09 11 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 +401 val_401 401 val_401 2008-04-09 12 402 val_402 402 val_402 2008-04-08 11 -402 val_402 402 val_402 2008-04-09 12 402 val_402 402 val_402 2008-04-08 12 +402 val_402 402 val_402 2008-04-09 11 +402 val_402 402 val_402 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 12 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 -403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-08 11 -403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-08 12 403 val_403 403 val_403 2008-04-08 12 403 val_403 403 val_403 2008-04-09 11 403 val_403 403 val_403 2008-04-09 11 -403 val_403 403 val_403 2008-04-08 12 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 11 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 +403 val_403 403 val_403 2008-04-09 12 403 val_403 403 val_403 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 11 -404 val_404 404 val_404 2008-04-08 12 -404 val_404 404 val_404 2008-04-09 11 -404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-08 12 -404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 11 404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-09 11 +404 val_404 404 val_404 2008-04-09 11 +404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-09 11 404 val_404 404 val_404 2008-04-09 12 -404 val_404 404 val_404 2008-04-08 12 +404 val_404 404 val_404 2008-04-09 12 +404 val_404 404 val_404 2008-04-09 12 +404 val_404 404 val_404 2008-04-09 12 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 11 406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 11 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 +406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 -406 val_406 406 val_406 2008-04-08 12 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 11 406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-09 12 406 val_406 406 val_406 2008-04-09 12 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 11 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 -406 val_406 406 val_406 2008-04-08 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 +406 val_406 406 val_406 2008-04-09 12 407 val_407 407 val_407 2008-04-08 11 407 val_407 407 val_407 2008-04-08 12 -407 val_407 407 val_407 2008-04-09 12 407 val_407 407 val_407 2008-04-09 11 +407 val_407 407 val_407 2008-04-09 12 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-08 11 +409 val_409 409 val_409 2008-04-08 11 +409 val_409 409 val_409 2008-04-08 11 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 +409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-08 11 -409 val_409 409 val_409 2008-04-08 12 409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-08 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 11 409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-08 11 409 val_409 409 val_409 2008-04-09 12 409 val_409 409 val_409 2008-04-09 12 -409 val_409 409 val_409 2008-04-08 12 -409 val_409 409 val_409 2008-04-09 11 -409 val_409 409 val_409 2008-04-09 11 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +409 val_409 409 val_409 2008-04-09 12 +41 val_41 41 val_41 2008-04-08 11 +41 val_41 41 val_41 2008-04-08 12 +41 val_41 41 val_41 2008-04-09 11 +41 val_41 41 val_41 2008-04-09 12 411 val_411 411 val_411 2008-04-08 11 411 val_411 411 val_411 2008-04-08 12 -411 val_411 411 val_411 2008-04-09 12 411 val_411 411 val_411 2008-04-09 11 -413 val_413 413 val_413 2008-04-09 11 -413 val_413 413 val_413 2008-04-08 12 -413 val_413 413 val_413 2008-04-09 11 -413 val_413 413 val_413 2008-04-09 12 -413 val_413 413 val_413 2008-04-08 12 +411 val_411 411 val_411 2008-04-09 12 413 val_413 413 val_413 2008-04-08 11 -413 val_413 413 val_413 2008-04-09 12 413 val_413 413 val_413 2008-04-08 11 -413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 11 +413 val_413 413 val_413 2008-04-08 12 +413 val_413 413 val_413 2008-04-08 12 +413 val_413 413 val_413 2008-04-08 12 413 val_413 413 val_413 2008-04-08 12 413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-09 11 +413 val_413 413 val_413 2008-04-09 12 +413 val_413 413 val_413 2008-04-09 12 413 val_413 413 val_413 2008-04-09 12 -413 val_413 413 val_413 2008-04-08 12 -413 val_413 413 val_413 2008-04-08 11 413 val_413 413 val_413 2008-04-09 12 -413 val_413 413 val_413 2008-04-08 11 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-08 12 -414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-09 12 -414 val_414 414 val_414 2008-04-08 12 -414 val_414 414 val_414 2008-04-09 11 -414 val_414 414 val_414 2008-04-09 12 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-08 12 -414 val_414 414 val_414 2008-04-09 11 414 val_414 414 val_414 2008-04-08 11 -414 val_414 414 val_414 2008-04-09 12 +414 val_414 414 val_414 2008-04-08 12 +414 val_414 414 val_414 2008-04-08 12 +414 val_414 414 val_414 2008-04-08 12 414 val_414 414 val_414 2008-04-08 12 414 val_414 414 val_414 2008-04-09 11 +414 val_414 414 val_414 2008-04-09 11 +414 val_414 414 val_414 2008-04-09 11 +414 val_414 414 val_414 2008-04-09 11 +414 val_414 414 val_414 2008-04-09 12 +414 val_414 414 val_414 2008-04-09 12 +414 val_414 414 val_414 2008-04-09 12 414 val_414 414 val_414 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 11 417 val_417 417 val_417 2008-04-08 12 -417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 +417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-08 12 417 val_417 417 val_417 2008-04-09 11 -417 val_417 417 val_417 2008-04-08 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 11 417 val_417 417 val_417 2008-04-09 12 417 val_417 417 val_417 2008-04-09 12 -417 val_417 417 val_417 2008-04-08 11 -417 val_417 417 val_417 2008-04-09 11 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +417 val_417 417 val_417 2008-04-09 12 +418 val_418 418 val_418 2008-04-08 11 418 val_418 418 val_418 2008-04-08 12 -418 val_418 418 val_418 2008-04-09 12 418 val_418 418 val_418 2008-04-09 11 -418 val_418 418 val_418 2008-04-08 11 -419 val_419 419 val_419 2008-04-08 12 +418 val_418 418 val_418 2008-04-09 12 419 val_419 419 val_419 2008-04-08 11 -419 val_419 419 val_419 2008-04-09 12 +419 val_419 419 val_419 2008-04-08 12 419 val_419 419 val_419 2008-04-09 11 +419 val_419 419 val_419 2008-04-09 12 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 11 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-08 12 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 11 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 +42 val_42 42 val_42 2008-04-09 12 421 val_421 421 val_421 2008-04-08 11 421 val_421 421 val_421 2008-04-08 12 421 val_421 421 val_421 2008-04-09 11 421 val_421 421 val_421 2008-04-09 12 424 val_424 424 val_424 2008-04-08 11 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-08 12 -424 val_424 424 val_424 2008-04-08 12 -424 val_424 424 val_424 2008-04-09 12 424 val_424 424 val_424 2008-04-08 11 424 val_424 424 val_424 2008-04-08 11 -424 val_424 424 val_424 2008-04-09 11 -424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-08 11 424 val_424 424 val_424 2008-04-08 12 424 val_424 424 val_424 2008-04-08 12 +424 val_424 424 val_424 2008-04-08 12 +424 val_424 424 val_424 2008-04-08 12 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 +424 val_424 424 val_424 2008-04-09 11 424 val_424 424 val_424 2008-04-09 12 -424 val_424 424 val_424 2008-04-08 11 -427 val_427 427 val_427 2008-04-09 12 +424 val_424 424 val_424 2008-04-09 12 +424 val_424 424 val_424 2008-04-09 12 +424 val_424 424 val_424 2008-04-09 12 +427 val_427 427 val_427 2008-04-08 11 427 val_427 427 val_427 2008-04-08 12 427 val_427 427 val_427 2008-04-09 11 -427 val_427 427 val_427 2008-04-08 11 -429 val_429 429 val_429 2008-04-09 11 +427 val_427 427 val_427 2008-04-09 12 +429 val_429 429 val_429 2008-04-08 11 +429 val_429 429 val_429 2008-04-08 11 429 val_429 429 val_429 2008-04-08 11 -429 val_429 429 val_429 2008-04-08 12 -429 val_429 429 val_429 2008-04-09 12 429 val_429 429 val_429 2008-04-08 11 -429 val_429 429 val_429 2008-04-09 12 +429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-08 12 429 val_429 429 val_429 2008-04-08 12 429 val_429 429 val_429 2008-04-09 11 429 val_429 429 val_429 2008-04-09 11 -429 val_429 429 val_429 2008-04-08 11 -429 val_429 429 val_429 2008-04-08 12 +429 val_429 429 val_429 2008-04-09 11 +429 val_429 429 val_429 2008-04-09 11 429 val_429 429 val_429 2008-04-09 12 -429 val_429 429 val_429 2008-04-08 11 429 val_429 429 val_429 2008-04-09 12 -429 val_429 429 val_429 2008-04-08 12 -429 val_429 429 val_429 2008-04-09 11 +429 val_429 429 val_429 2008-04-09 12 +429 val_429 429 val_429 2008-04-09 12 +43 val_43 43 val_43 2008-04-08 11 +43 val_43 43 val_43 2008-04-08 12 +43 val_43 43 val_43 2008-04-09 11 +43 val_43 43 val_43 2008-04-09 12 +430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 11 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 11 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-08 12 -430 val_430 430 val_430 2008-04-08 11 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-08 12 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-09 11 430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 11 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 +430 val_430 430 val_430 2008-04-09 12 430 val_430 430 val_430 2008-04-09 12 -430 val_430 430 val_430 2008-04-08 11 -430 val_430 430 val_430 2008-04-08 12 430 val_430 430 val_430 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-08 11 431 val_431 431 val_431 2008-04-08 11 431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-08 12 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-08 11 431 val_431 431 val_431 2008-04-08 11 431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 -431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 11 +431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-08 12 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 11 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 +431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 431 val_431 431 val_431 2008-04-09 12 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -431 val_431 431 val_431 2008-04-08 11 -432 val_432 432 val_432 2008-04-09 12 432 val_432 432 val_432 2008-04-08 11 -432 val_432 432 val_432 2008-04-09 11 432 val_432 432 val_432 2008-04-08 12 -435 val_435 435 val_435 2008-04-09 11 +432 val_432 432 val_432 2008-04-09 11 +432 val_432 432 val_432 2008-04-09 12 435 val_435 435 val_435 2008-04-08 11 435 val_435 435 val_435 2008-04-08 12 +435 val_435 435 val_435 2008-04-09 11 435 val_435 435 val_435 2008-04-09 12 -436 val_436 436 val_436 2008-04-08 12 436 val_436 436 val_436 2008-04-08 11 +436 val_436 436 val_436 2008-04-08 12 436 val_436 436 val_436 2008-04-09 11 436 val_436 436 val_436 2008-04-09 12 -437 val_437 437 val_437 2008-04-09 12 437 val_437 437 val_437 2008-04-08 11 -437 val_437 437 val_437 2008-04-09 11 437 val_437 437 val_437 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 +437 val_437 437 val_437 2008-04-09 11 +437 val_437 437 val_437 2008-04-09 12 +438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 12 -438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-08 11 -438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-08 11 438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 11 438 val_438 438 val_438 2008-04-08 12 -438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 +438 val_438 438 val_438 2008-04-08 12 438 val_438 438 val_438 2008-04-09 11 -438 val_438 438 val_438 2008-04-08 11 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-08 12 -439 val_439 439 val_439 2008-04-08 12 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 11 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +438 val_438 438 val_438 2008-04-09 12 +439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-08 11 439 val_439 439 val_439 2008-04-08 11 -439 val_439 439 val_439 2008-04-09 12 439 val_439 439 val_439 2008-04-08 11 -439 val_439 439 val_439 2008-04-09 12 -439 val_439 439 val_439 2008-04-09 11 -439 val_439 439 val_439 2008-04-09 11 439 val_439 439 val_439 2008-04-08 12 439 val_439 439 val_439 2008-04-08 12 -439 val_439 439 val_439 2008-04-08 11 +439 val_439 439 val_439 2008-04-08 12 +439 val_439 439 val_439 2008-04-08 12 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 +439 val_439 439 val_439 2008-04-09 11 439 val_439 439 val_439 2008-04-09 12 -439 val_439 439 val_439 2008-04-08 11 439 val_439 439 val_439 2008-04-09 12 -443 val_443 443 val_443 2008-04-09 11 -443 val_443 443 val_443 2008-04-09 12 +439 val_439 439 val_439 2008-04-09 12 +439 val_439 439 val_439 2008-04-09 12 +44 val_44 44 val_44 2008-04-08 11 +44 val_44 44 val_44 2008-04-08 12 +44 val_44 44 val_44 2008-04-09 11 +44 val_44 44 val_44 2008-04-09 12 443 val_443 443 val_443 2008-04-08 11 443 val_443 443 val_443 2008-04-08 12 -444 val_444 444 val_444 2008-04-09 11 +443 val_443 443 val_443 2008-04-09 11 +443 val_443 443 val_443 2008-04-09 12 444 val_444 444 val_444 2008-04-08 11 444 val_444 444 val_444 2008-04-08 12 +444 val_444 444 val_444 2008-04-09 11 444 val_444 444 val_444 2008-04-09 12 -446 val_446 446 val_446 2008-04-08 12 -446 val_446 446 val_446 2008-04-09 12 446 val_446 446 val_446 2008-04-08 11 +446 val_446 446 val_446 2008-04-08 12 446 val_446 446 val_446 2008-04-09 11 -448 val_448 448 val_448 2008-04-09 11 -448 val_448 448 val_448 2008-04-09 12 +446 val_446 446 val_446 2008-04-09 12 448 val_448 448 val_448 2008-04-08 11 448 val_448 448 val_448 2008-04-08 12 -449 val_449 449 val_449 2008-04-08 12 +448 val_448 448 val_448 2008-04-09 11 +448 val_448 448 val_448 2008-04-09 12 449 val_449 449 val_449 2008-04-08 11 -449 val_449 449 val_449 2008-04-09 12 +449 val_449 449 val_449 2008-04-08 12 449 val_449 449 val_449 2008-04-09 11 +449 val_449 449 val_449 2008-04-09 12 452 val_452 452 val_452 2008-04-08 11 +452 val_452 452 val_452 2008-04-08 12 452 val_452 452 val_452 2008-04-09 11 452 val_452 452 val_452 2008-04-09 12 -452 val_452 452 val_452 2008-04-08 12 +453 val_453 453 val_453 2008-04-08 11 453 val_453 453 val_453 2008-04-08 12 453 val_453 453 val_453 2008-04-09 11 453 val_453 453 val_453 2008-04-09 12 -453 val_453 453 val_453 2008-04-08 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-08 11 +454 val_454 454 val_454 2008-04-08 11 +454 val_454 454 val_454 2008-04-08 11 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-08 12 454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-08 12 454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 11 -454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-08 11 454 val_454 454 val_454 2008-04-09 11 -454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 +454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-09 11 454 val_454 454 val_454 2008-04-09 12 454 val_454 454 val_454 2008-04-09 12 -454 val_454 454 val_454 2008-04-08 12 -454 val_454 454 val_454 2008-04-08 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 +454 val_454 454 val_454 2008-04-09 12 455 val_455 455 val_455 2008-04-08 11 -455 val_455 455 val_455 2008-04-09 11 455 val_455 455 val_455 2008-04-08 12 +455 val_455 455 val_455 2008-04-09 11 455 val_455 455 val_455 2008-04-09 12 -457 val_457 457 val_457 2008-04-08 12 -457 val_457 457 val_457 2008-04-09 12 457 val_457 457 val_457 2008-04-08 11 +457 val_457 457 val_457 2008-04-08 12 457 val_457 457 val_457 2008-04-09 11 -458 val_458 458 val_458 2008-04-08 12 -458 val_458 458 val_458 2008-04-09 11 +457 val_457 457 val_457 2008-04-09 12 458 val_458 458 val_458 2008-04-08 11 -458 val_458 458 val_458 2008-04-09 11 458 val_458 458 val_458 2008-04-08 11 -458 val_458 458 val_458 2008-04-09 12 -458 val_458 458 val_458 2008-04-09 12 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-08 12 +458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-08 12 458 val_458 458 val_458 2008-04-09 11 -458 val_458 458 val_458 2008-04-08 11 458 val_458 458 val_458 2008-04-09 11 -458 val_458 458 val_458 2008-04-08 11 +458 val_458 458 val_458 2008-04-09 11 +458 val_458 458 val_458 2008-04-09 11 +458 val_458 458 val_458 2008-04-09 12 +458 val_458 458 val_458 2008-04-09 12 458 val_458 458 val_458 2008-04-09 12 458 val_458 458 val_458 2008-04-09 12 -458 val_458 458 val_458 2008-04-08 12 -459 val_459 459 val_459 2008-04-08 12 -459 val_459 459 val_459 2008-04-09 12 -459 val_459 459 val_459 2008-04-09 11 -459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-08 11 -459 val_459 459 val_459 2008-04-08 12 -459 val_459 459 val_459 2008-04-09 12 +459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-08 11 459 val_459 459 val_459 2008-04-08 11 459 val_459 459 val_459 2008-04-08 12 -459 val_459 459 val_459 2008-04-09 12 +459 val_459 459 val_459 2008-04-08 12 +459 val_459 459 val_459 2008-04-08 12 +459 val_459 459 val_459 2008-04-08 12 +459 val_459 459 val_459 2008-04-09 11 +459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-09 11 459 val_459 459 val_459 2008-04-09 11 -459 val_459 459 val_459 2008-04-08 11 -459 val_459 459 val_459 2008-04-08 12 459 val_459 459 val_459 2008-04-09 12 -459 val_459 459 val_459 2008-04-08 11 +459 val_459 459 val_459 2008-04-09 12 +459 val_459 459 val_459 2008-04-09 12 +459 val_459 459 val_459 2008-04-09 12 460 val_460 460 val_460 2008-04-08 11 -460 val_460 460 val_460 2008-04-09 11 460 val_460 460 val_460 2008-04-08 12 +460 val_460 460 val_460 2008-04-09 11 460 val_460 460 val_460 2008-04-09 12 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 12 462 val_462 462 val_462 2008-04-08 11 462 val_462 462 val_462 2008-04-08 11 -462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-08 11 +462 val_462 462 val_462 2008-04-08 11 +462 val_462 462 val_462 2008-04-08 12 +462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-08 12 462 val_462 462 val_462 2008-04-09 11 -462 val_462 462 val_462 2008-04-09 12 -462 val_462 462 val_462 2008-04-08 11 -462 val_462 462 val_462 2008-04-08 11 +462 val_462 462 val_462 2008-04-09 11 +462 val_462 462 val_462 2008-04-09 11 462 val_462 462 val_462 2008-04-09 11 462 val_462 462 val_462 2008-04-09 12 -462 val_462 462 val_462 2008-04-08 12 -462 val_462 462 val_462 2008-04-08 12 -463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-08 12 +462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-09 12 +462 val_462 462 val_462 2008-04-09 12 +463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-08 11 463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-08 12 +463 val_463 463 val_463 2008-04-08 12 +463 val_463 463 val_463 2008-04-08 12 463 val_463 463 val_463 2008-04-08 12 463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-09 11 +463 val_463 463 val_463 2008-04-09 11 463 val_463 463 val_463 2008-04-09 11 463 val_463 463 val_463 2008-04-09 12 463 val_463 463 val_463 2008-04-09 12 -463 val_463 463 val_463 2008-04-08 12 -463 val_463 463 val_463 2008-04-08 11 -463 val_463 463 val_463 2008-04-08 12 -463 val_463 463 val_463 2008-04-09 11 -463 val_463 463 val_463 2008-04-08 11 +463 val_463 463 val_463 2008-04-09 12 +463 val_463 463 val_463 2008-04-09 12 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-08 11 466 val_466 466 val_466 2008-04-08 11 -466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 +466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-09 11 -466 val_466 466 val_466 2008-04-08 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 +466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 11 466 val_466 466 val_466 2008-04-09 12 466 val_466 466 val_466 2008-04-09 12 -466 val_466 466 val_466 2008-04-08 12 -466 val_466 466 val_466 2008-04-08 12 466 val_466 466 val_466 2008-04-09 12 -467 val_467 467 val_467 2008-04-09 12 -467 val_467 467 val_467 2008-04-09 11 -467 val_467 467 val_467 2008-04-08 12 +466 val_466 466 val_466 2008-04-09 12 +466 val_466 466 val_466 2008-04-09 12 +466 val_466 466 val_466 2008-04-09 12 +466 val_466 466 val_466 2008-04-09 12 +466 val_466 466 val_466 2008-04-09 12 +466 val_466 466 val_466 2008-04-09 12 467 val_467 467 val_467 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 12 +467 val_467 467 val_467 2008-04-08 12 +467 val_467 467 val_467 2008-04-09 11 +467 val_467 467 val_467 2008-04-09 12 +468 val_468 468 val_468 2008-04-08 11 +468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-08 12 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 11 -468 val_468 468 val_468 2008-04-09 12 468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 +468 val_468 468 val_468 2008-04-08 12 468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-09 11 -468 val_468 468 val_468 2008-04-09 12 -468 val_468 468 val_468 2008-04-08 11 468 val_468 468 val_468 2008-04-09 11 468 val_468 468 val_468 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 11 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +468 val_468 468 val_468 2008-04-09 12 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-08 12 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 11 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 +469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-09 11 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 12 -469 val_469 469 val_469 2008-04-08 11 469 val_469 469 val_469 2008-04-09 12 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-09 11 469 val_469 469 val_469 2008-04-09 12 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 11 -469 val_469 469 val_469 2008-04-08 12 469 val_469 469 val_469 2008-04-09 12 +47 val_47 47 val_47 2008-04-08 11 +47 val_47 47 val_47 2008-04-08 12 +47 val_47 47 val_47 2008-04-09 11 +47 val_47 47 val_47 2008-04-09 12 470 val_470 470 val_470 2008-04-08 11 -470 val_470 470 val_470 2008-04-09 12 470 val_470 470 val_470 2008-04-08 12 470 val_470 470 val_470 2008-04-09 11 -472 val_472 472 val_472 2008-04-09 11 +470 val_470 470 val_470 2008-04-09 12 +472 val_472 472 val_472 2008-04-08 11 472 val_472 472 val_472 2008-04-08 12 +472 val_472 472 val_472 2008-04-09 11 472 val_472 472 val_472 2008-04-09 12 -472 val_472 472 val_472 2008-04-08 11 -475 val_475 475 val_475 2008-04-09 11 475 val_475 475 val_475 2008-04-08 11 475 val_475 475 val_475 2008-04-08 12 +475 val_475 475 val_475 2008-04-09 11 475 val_475 475 val_475 2008-04-09 12 -477 val_477 477 val_477 2008-04-09 12 477 val_477 477 val_477 2008-04-08 11 477 val_477 477 val_477 2008-04-08 12 477 val_477 477 val_477 2008-04-09 11 +477 val_477 477 val_477 2008-04-09 12 478 val_478 478 val_478 2008-04-08 11 -478 val_478 478 val_478 2008-04-08 12 -478 val_478 478 val_478 2008-04-09 11 -478 val_478 478 val_478 2008-04-09 12 -478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-08 11 -478 val_478 478 val_478 2008-04-09 11 -478 val_478 478 val_478 2008-04-08 12 478 val_478 478 val_478 2008-04-08 11 +478 val_478 478 val_478 2008-04-08 11 +478 val_478 478 val_478 2008-04-08 12 +478 val_478 478 val_478 2008-04-08 12 +478 val_478 478 val_478 2008-04-08 12 478 val_478 478 val_478 2008-04-08 12 478 val_478 478 val_478 2008-04-09 11 +478 val_478 478 val_478 2008-04-09 11 +478 val_478 478 val_478 2008-04-09 11 +478 val_478 478 val_478 2008-04-09 11 +478 val_478 478 val_478 2008-04-09 12 +478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-09 12 478 val_478 478 val_478 2008-04-09 12 -478 val_478 478 val_478 2008-04-08 11 -478 val_478 478 val_478 2008-04-09 11 -478 val_478 478 val_478 2008-04-08 12 -479 val_479 479 val_479 2008-04-08 12 479 val_479 479 val_479 2008-04-08 11 +479 val_479 479 val_479 2008-04-08 12 479 val_479 479 val_479 2008-04-09 11 479 val_479 479 val_479 2008-04-09 12 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 480 val_480 480 val_480 2008-04-08 11 -480 val_480 480 val_480 2008-04-09 12 480 val_480 480 val_480 2008-04-08 12 -480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 +480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-08 12 480 val_480 480 val_480 2008-04-09 11 480 val_480 480 val_480 2008-04-09 11 480 val_480 480 val_480 2008-04-09 11 -480 val_480 480 val_480 2008-04-08 12 -481 val_481 481 val_481 2008-04-08 12 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 11 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 +480 val_480 480 val_480 2008-04-09 12 481 val_481 481 val_481 2008-04-08 11 +481 val_481 481 val_481 2008-04-08 12 481 val_481 481 val_481 2008-04-09 11 481 val_481 481 val_481 2008-04-09 12 -482 val_482 482 val_482 2008-04-09 11 -482 val_482 482 val_482 2008-04-09 12 482 val_482 482 val_482 2008-04-08 11 482 val_482 482 val_482 2008-04-08 12 -483 val_483 483 val_483 2008-04-09 11 +482 val_482 482 val_482 2008-04-09 11 +482 val_482 482 val_482 2008-04-09 12 483 val_483 483 val_483 2008-04-08 11 -483 val_483 483 val_483 2008-04-09 12 483 val_483 483 val_483 2008-04-08 12 -484 val_484 484 val_484 2008-04-08 12 +483 val_483 483 val_483 2008-04-09 11 +483 val_483 483 val_483 2008-04-09 12 484 val_484 484 val_484 2008-04-08 11 +484 val_484 484 val_484 2008-04-08 12 484 val_484 484 val_484 2008-04-09 11 484 val_484 484 val_484 2008-04-09 12 485 val_485 485 val_485 2008-04-08 11 +485 val_485 485 val_485 2008-04-08 12 485 val_485 485 val_485 2008-04-09 11 485 val_485 485 val_485 2008-04-09 12 -485 val_485 485 val_485 2008-04-08 12 -487 val_487 487 val_487 2008-04-09 12 +487 val_487 487 val_487 2008-04-08 11 487 val_487 487 val_487 2008-04-08 12 487 val_487 487 val_487 2008-04-09 11 -487 val_487 487 val_487 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 +487 val_487 487 val_487 2008-04-09 12 +489 val_489 489 val_489 2008-04-08 11 +489 val_489 489 val_489 2008-04-08 11 +489 val_489 489 val_489 2008-04-08 11 +489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-09 12 489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 +489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-09 11 489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 489 val_489 489 val_489 2008-04-09 11 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-08 11 -489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 11 489 val_489 489 val_489 2008-04-09 11 489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 489 val_489 489 val_489 2008-04-09 12 -489 val_489 489 val_489 2008-04-08 12 -489 val_489 489 val_489 2008-04-09 11 -490 val_490 490 val_490 2008-04-09 11 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 +489 val_489 489 val_489 2008-04-09 12 490 val_490 490 val_490 2008-04-08 11 -490 val_490 490 val_490 2008-04-09 12 490 val_490 490 val_490 2008-04-08 12 +490 val_490 490 val_490 2008-04-09 11 +490 val_490 490 val_490 2008-04-09 12 491 val_491 491 val_491 2008-04-08 11 491 val_491 491 val_491 2008-04-08 12 491 val_491 491 val_491 2008-04-09 11 491 val_491 491 val_491 2008-04-09 12 492 val_492 492 val_492 2008-04-08 11 -492 val_492 492 val_492 2008-04-09 11 -492 val_492 492 val_492 2008-04-09 12 -492 val_492 492 val_492 2008-04-09 12 -492 val_492 492 val_492 2008-04-08 12 -492 val_492 492 val_492 2008-04-09 11 492 val_492 492 val_492 2008-04-08 11 -492 val_492 492 val_492 2008-04-08 12 492 val_492 492 val_492 2008-04-08 11 +492 val_492 492 val_492 2008-04-08 11 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-08 12 +492 val_492 492 val_492 2008-04-09 11 492 val_492 492 val_492 2008-04-09 11 +492 val_492 492 val_492 2008-04-09 11 +492 val_492 492 val_492 2008-04-09 11 +492 val_492 492 val_492 2008-04-09 12 +492 val_492 492 val_492 2008-04-09 12 492 val_492 492 val_492 2008-04-09 12 492 val_492 492 val_492 2008-04-09 12 -492 val_492 492 val_492 2008-04-08 12 -492 val_492 492 val_492 2008-04-09 11 -492 val_492 492 val_492 2008-04-08 11 -492 val_492 492 val_492 2008-04-08 12 +493 val_493 493 val_493 2008-04-08 11 493 val_493 493 val_493 2008-04-08 12 -493 val_493 493 val_493 2008-04-09 12 493 val_493 493 val_493 2008-04-09 11 -493 val_493 493 val_493 2008-04-08 11 -494 val_494 494 val_494 2008-04-09 11 +493 val_493 493 val_493 2008-04-09 12 494 val_494 494 val_494 2008-04-08 11 494 val_494 494 val_494 2008-04-08 12 +494 val_494 494 val_494 2008-04-09 11 494 val_494 494 val_494 2008-04-09 12 495 val_495 495 val_495 2008-04-08 11 -495 val_495 495 val_495 2008-04-09 12 495 val_495 495 val_495 2008-04-08 12 495 val_495 495 val_495 2008-04-09 11 +495 val_495 495 val_495 2008-04-09 12 +496 val_496 496 val_496 2008-04-08 11 +496 val_496 496 val_496 2008-04-08 12 496 val_496 496 val_496 2008-04-09 11 496 val_496 496 val_496 2008-04-09 12 -496 val_496 496 val_496 2008-04-08 12 -496 val_496 496 val_496 2008-04-08 11 -497 val_497 497 val_497 2008-04-09 12 -497 val_497 497 val_497 2008-04-08 12 497 val_497 497 val_497 2008-04-08 11 +497 val_497 497 val_497 2008-04-08 12 497 val_497 497 val_497 2008-04-09 11 -498 val_498 498 val_498 2008-04-08 12 +497 val_497 497 val_497 2008-04-09 12 +498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 12 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-08 11 498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-08 12 498 val_498 498 val_498 2008-04-08 12 -498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 11 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-09 11 +498 val_498 498 val_498 2008-04-09 11 +498 val_498 498 val_498 2008-04-09 11 +498 val_498 498 val_498 2008-04-09 11 +498 val_498 498 val_498 2008-04-09 11 +498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 11 -498 val_498 498 val_498 2008-04-08 11 -498 val_498 498 val_498 2008-04-08 12 498 val_498 498 val_498 2008-04-09 12 498 val_498 498 val_498 2008-04-09 12 -498 val_498 498 val_498 2008-04-08 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +498 val_498 498 val_498 2008-04-09 12 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 11 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-08 12 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 11 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +5 val_5 5 val_5 2008-04-09 12 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 11 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-08 12 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 11 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +51 val_51 51 val_51 2008-04-09 12 +53 val_53 53 val_53 2008-04-08 11 +53 val_53 53 val_53 2008-04-08 12 +53 val_53 53 val_53 2008-04-09 11 +53 val_53 53 val_53 2008-04-09 12 +54 val_54 54 val_54 2008-04-08 11 +54 val_54 54 val_54 2008-04-08 12 +54 val_54 54 val_54 2008-04-09 11 +54 val_54 54 val_54 2008-04-09 12 +57 val_57 57 val_57 2008-04-08 11 +57 val_57 57 val_57 2008-04-08 12 +57 val_57 57 val_57 2008-04-09 11 +57 val_57 57 val_57 2008-04-09 12 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 11 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-08 12 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 11 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +58 val_58 58 val_58 2008-04-09 12 +64 val_64 64 val_64 2008-04-08 11 +64 val_64 64 val_64 2008-04-08 12 +64 val_64 64 val_64 2008-04-09 11 +64 val_64 64 val_64 2008-04-09 12 +65 val_65 65 val_65 2008-04-08 11 +65 val_65 65 val_65 2008-04-08 12 +65 val_65 65 val_65 2008-04-09 11 +65 val_65 65 val_65 2008-04-09 12 +66 val_66 66 val_66 2008-04-08 11 +66 val_66 66 val_66 2008-04-08 12 +66 val_66 66 val_66 2008-04-09 11 +66 val_66 66 val_66 2008-04-09 12 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 11 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-08 12 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 11 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +67 val_67 67 val_67 2008-04-09 12 +69 val_69 69 val_69 2008-04-08 11 +69 val_69 69 val_69 2008-04-08 12 +69 val_69 69 val_69 2008-04-09 11 +69 val_69 69 val_69 2008-04-09 12 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 11 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-08 12 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 11 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +70 val_70 70 val_70 2008-04-09 12 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 11 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-08 12 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 11 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +72 val_72 72 val_72 2008-04-09 12 +74 val_74 74 val_74 2008-04-08 11 +74 val_74 74 val_74 2008-04-08 12 +74 val_74 74 val_74 2008-04-09 11 +74 val_74 74 val_74 2008-04-09 12 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 11 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-08 12 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 11 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +76 val_76 76 val_76 2008-04-09 12 +77 val_77 77 val_77 2008-04-08 11 +77 val_77 77 val_77 2008-04-08 12 +77 val_77 77 val_77 2008-04-09 11 +77 val_77 77 val_77 2008-04-09 12 +78 val_78 78 val_78 2008-04-08 11 +78 val_78 78 val_78 2008-04-08 12 +78 val_78 78 val_78 2008-04-09 11 +78 val_78 78 val_78 2008-04-09 12 +8 val_8 8 val_8 2008-04-08 11 +8 val_8 8 val_8 2008-04-08 12 +8 val_8 8 val_8 2008-04-09 11 +8 val_8 8 val_8 2008-04-09 12 +80 val_80 80 val_80 2008-04-08 11 +80 val_80 80 val_80 2008-04-08 12 +80 val_80 80 val_80 2008-04-09 11 +80 val_80 80 val_80 2008-04-09 12 +82 val_82 82 val_82 2008-04-08 11 +82 val_82 82 val_82 2008-04-08 12 +82 val_82 82 val_82 2008-04-09 11 +82 val_82 82 val_82 2008-04-09 12 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 11 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-08 12 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 11 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +83 val_83 83 val_83 2008-04-09 12 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 11 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-08 12 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 11 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +84 val_84 84 val_84 2008-04-09 12 +85 val_85 85 val_85 2008-04-08 11 +85 val_85 85 val_85 2008-04-08 12 +85 val_85 85 val_85 2008-04-09 11 +85 val_85 85 val_85 2008-04-09 12 +86 val_86 86 val_86 2008-04-08 11 +86 val_86 86 val_86 2008-04-08 12 +86 val_86 86 val_86 2008-04-09 11 +86 val_86 86 val_86 2008-04-09 12 +87 val_87 87 val_87 2008-04-08 11 +87 val_87 87 val_87 2008-04-08 12 +87 val_87 87 val_87 2008-04-09 11 +87 val_87 87 val_87 2008-04-09 12 +9 val_9 9 val_9 2008-04-08 11 +9 val_9 9 val_9 2008-04-08 12 +9 val_9 9 val_9 2008-04-09 11 +9 val_9 9 val_9 2008-04-09 12 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 11 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-08 12 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 11 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +90 val_90 90 val_90 2008-04-09 12 +92 val_92 92 val_92 2008-04-08 11 +92 val_92 92 val_92 2008-04-08 12 +92 val_92 92 val_92 2008-04-09 11 +92 val_92 92 val_92 2008-04-09 12 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 11 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-08 12 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 11 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +95 val_95 95 val_95 2008-04-09 12 +96 val_96 96 val_96 2008-04-08 11 +96 val_96 96 val_96 2008-04-08 12 +96 val_96 96 val_96 2008-04-09 11 +96 val_96 96 val_96 2008-04-09 12 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 11 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-08 12 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 11 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +97 val_97 97 val_97 2008-04-09 12 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 11 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-08 12 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 11 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 +98 val_98 98 val_98 2008-04-09 12 PREHOOK: query: explain select * from default.masking_test_subq where key > 0 PREHOOK: type: QUERY PREHOOK: Input: default@masking_test_subq @@ -6639,159 +6704,93 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from default.masking_test_subq where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq -PREHOOK: Input: default@src + Tez #### A masked pattern was here #### -POSTHOOK: query: select * from default.masking_test_subq where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq -POSTHOOK: Input: default@src + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) #### A masked pattern was here #### -2 val_2 -4 val_4 -5 val_5 -5 val_5 -5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 -51 val_51 -51 val_51 -53 val_53 -54 val_54 -57 val_57 -58 val_58 -58 val_58 -64 val_64 -65 val_65 -66 val_66 -67 val_67 -67 val_67 -69 val_69 -70 val_70 -70 val_70 -70 val_70 -72 val_72 -72 val_72 -74 val_74 -76 val_76 -76 val_76 -77 val_77 -78 val_78 -80 val_80 -82 val_82 -83 val_83 -83 val_83 -84 val_84 -84 val_84 -85 val_85 -86 val_86 -87 val_87 -90 val_90 -90 val_90 -90 val_90 -92 val_92 -95 val_95 -95 val_95 -96 val_96 -97 val_97 -97 val_97 -98 val_98 -98 val_98 + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from default.masking_test_subq where key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from default.masking_test_subq where key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 100 val_100 100 val_100 103 val_103 @@ -6799,6 +6798,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -6809,6 +6809,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -6836,6 +6838,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -6860,6 +6864,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -6873,6 +6878,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -6881,6 +6888,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -6897,6 +6905,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -6947,6 +6957,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -6961,12 +6973,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -6980,6 +6995,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -7000,6 +7016,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -7031,6 +7048,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -7040,6 +7058,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -7051,6 +7070,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -7066,6 +7088,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -7091,6 +7115,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -7111,6 +7136,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -7121,12 +7147,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -7142,6 +7171,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -7176,6 +7206,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -7208,132 +7239,9 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 -PREHOOK: query: explain select * from masking_test_subq where masking_test_subq.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_subq where masking_test_subq.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq - filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_subq where masking_test_subq.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_subq where masking_test_subq.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq -POSTHOOK: Input: default@src -#### A masked pattern was here #### -2 val_2 -4 val_4 5 val_5 5 val_5 5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -20 val_20 -24 val_24 -24 val_24 -26 val_26 -26 val_26 -27 val_27 -28 val_28 -30 val_30 -33 val_33 -34 val_34 -35 val_35 -35 val_35 -35 val_35 -37 val_37 -37 val_37 -41 val_41 -42 val_42 -42 val_42 -43 val_43 -44 val_44 -47 val_47 51 val_51 51 val_51 53 val_53 @@ -7357,6 +7265,7 @@ POSTHOOK: Input: default@src 76 val_76 77 val_77 78 val_78 +8 val_8 80 val_80 82 val_82 83 val_83 @@ -7366,6 +7275,7 @@ POSTHOOK: Input: default@src 85 val_85 86 val_86 87 val_87 +9 val_9 90 val_90 90 val_90 90 val_90 @@ -7377,6 +7287,109 @@ POSTHOOK: Input: default@src 97 val_97 98 val_98 98 val_98 +PREHOOK: query: explain select * from masking_test_subq where masking_test_subq.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from masking_test_subq where masking_test_subq.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq + filterExpr: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: (UDFToDouble(key) is not null and (key > 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select * from masking_test_subq where masking_test_subq.key > 0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: select * from masking_test_subq where masking_test_subq.key > 0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq +POSTHOOK: Input: default@src +#### A masked pattern was here #### +10 val_10 100 val_100 100 val_100 103 val_103 @@ -7384,6 +7397,7 @@ POSTHOOK: Input: default@src 104 val_104 104 val_104 105 val_105 +11 val_11 111 val_111 113 val_113 113 val_113 @@ -7394,6 +7408,8 @@ POSTHOOK: Input: default@src 119 val_119 119 val_119 119 val_119 +12 val_12 +12 val_12 120 val_120 120 val_120 125 val_125 @@ -7421,6 +7437,8 @@ POSTHOOK: Input: default@src 146 val_146 149 val_149 149 val_149 +15 val_15 +15 val_15 150 val_150 152 val_152 152 val_152 @@ -7445,6 +7463,7 @@ POSTHOOK: Input: default@src 169 val_169 169 val_169 169 val_169 +17 val_17 170 val_170 172 val_172 172 val_172 @@ -7458,6 +7477,8 @@ POSTHOOK: Input: default@src 178 val_178 179 val_179 179 val_179 +18 val_18 +18 val_18 180 val_180 181 val_181 183 val_183 @@ -7466,6 +7487,7 @@ POSTHOOK: Input: default@src 187 val_187 187 val_187 189 val_189 +19 val_19 190 val_190 191 val_191 191 val_191 @@ -7482,6 +7504,8 @@ POSTHOOK: Input: default@src 199 val_199 199 val_199 199 val_199 +2 val_2 +20 val_20 200 val_200 200 val_200 201 val_201 @@ -7532,6 +7556,8 @@ POSTHOOK: Input: default@src 238 val_238 239 val_239 239 val_239 +24 val_24 +24 val_24 241 val_241 242 val_242 242 val_242 @@ -7546,12 +7572,15 @@ POSTHOOK: Input: default@src 256 val_256 257 val_257 258 val_258 +26 val_26 +26 val_26 260 val_260 262 val_262 263 val_263 265 val_265 265 val_265 266 val_266 +27 val_27 272 val_272 272 val_272 273 val_273 @@ -7565,6 +7594,7 @@ POSTHOOK: Input: default@src 277 val_277 278 val_278 278 val_278 +28 val_28 280 val_280 280 val_280 281 val_281 @@ -7585,6 +7615,7 @@ POSTHOOK: Input: default@src 298 val_298 298 val_298 298 val_298 +30 val_30 302 val_302 305 val_305 306 val_306 @@ -7616,6 +7647,7 @@ POSTHOOK: Input: default@src 327 val_327 327 val_327 327 val_327 +33 val_33 331 val_331 331 val_331 332 val_332 @@ -7625,6 +7657,7 @@ POSTHOOK: Input: default@src 336 val_336 338 val_338 339 val_339 +34 val_34 341 val_341 342 val_342 342 val_342 @@ -7636,6 +7669,9 @@ POSTHOOK: Input: default@src 348 val_348 348 val_348 348 val_348 +35 val_35 +35 val_35 +35 val_35 351 val_351 353 val_353 353 val_353 @@ -7651,6 +7687,8 @@ POSTHOOK: Input: default@src 369 val_369 369 val_369 369 val_369 +37 val_37 +37 val_37 373 val_373 374 val_374 375 val_375 @@ -7676,6 +7714,7 @@ POSTHOOK: Input: default@src 397 val_397 399 val_399 399 val_399 +4 val_4 400 val_400 401 val_401 401 val_401 @@ -7696,6 +7735,7 @@ POSTHOOK: Input: default@src 409 val_409 409 val_409 409 val_409 +41 val_41 411 val_411 413 val_413 413 val_413 @@ -7706,12 +7746,15 @@ POSTHOOK: Input: default@src 417 val_417 418 val_418 419 val_419 +42 val_42 +42 val_42 421 val_421 424 val_424 424 val_424 427 val_427 429 val_429 429 val_429 +43 val_43 430 val_430 430 val_430 430 val_430 @@ -7727,6 +7770,7 @@ POSTHOOK: Input: default@src 438 val_438 439 val_439 439 val_439 +44 val_44 443 val_443 444 val_444 446 val_446 @@ -7761,6 +7805,7 @@ POSTHOOK: Input: default@src 469 val_469 469 val_469 469 val_469 +47 val_47 470 val_470 472 val_472 475 val_475 @@ -7793,6 +7838,54 @@ POSTHOOK: Input: default@src 498 val_498 498 val_498 498 val_498 +5 val_5 +5 val_5 +5 val_5 +51 val_51 +51 val_51 +53 val_53 +54 val_54 +57 val_57 +58 val_58 +58 val_58 +64 val_64 +65 val_65 +66 val_66 +67 val_67 +67 val_67 +69 val_69 +70 val_70 +70 val_70 +70 val_70 +72 val_72 +72 val_72 +74 val_74 +76 val_76 +76 val_76 +77 val_77 +78 val_78 +8 val_8 +80 val_80 +82 val_82 +83 val_83 +83 val_83 +84 val_84 +84 val_84 +85 val_85 +86 val_86 +87 val_87 +9 val_9 +90 val_90 +90 val_90 +90 val_90 +92 val_92 +95 val_95 +95 val_95 +96 val_96 +97 val_97 +97 val_97 +98 val_98 +98 val_98 PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -7802,36 +7895,20 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@src #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), upper(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: src + filterExpr: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) + Filter Operator + predicate: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) + Select Operator + expressions: key (type: string), upper(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink diff --git ql/src/test/results/clientpositive/llap/masking_disablecbo_4.q.out ql/src/test/results/clientpositive/llap/masking_disablecbo_4.q.out new file mode 100644 index 0000000000..47b47cd51f --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_disablecbo_4.q.out @@ -0,0 +1,212 @@ +PREHOOK: query: create table masking_test_n3 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n3 +POSTHOOK: query: create table masking_test_n3 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n3 +POSTHOOK: Lineage: masking_test_n3.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_n3.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: create table masking_test_subq_n0 as select cast(key as int) as key, value from src +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@src +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_subq_n0 +POSTHOOK: query: create table masking_test_subq_n0 as select cast(key as int) as key, value from src +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@src +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_subq_n0 +POSTHOOK: Lineage: masking_test_subq_n0.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: masking_test_subq_n0.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +with q1 as ( select key from q2 where key = '5'), +q2 as ( select key from src where key = '5') +select * from (select key from q1) a +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select key from q2 where key = '5'), +q2 as ( select key from src where key = '5') +select * from (select key from q1) a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: src + filterExpr: (key = '5') (type: boolean) + Filter Operator + predicate: (key = '5') (type: boolean) + Select Operator + expressions: '5' (type: string) + outputColumnNames: _col0 + ListSink + +PREHOOK: query: explain +with q1 as ( select * from masking_test_n3 where key = '5') +select * from q1 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n3 +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select * from masking_test_n3 where key = '5') +select * from q1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n3 + filterExpr: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) + Select Operator + expressions: 5 (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: explain +with masking_test_subq_n0 as ( select * from masking_test_n3 where key = '5') +select * from masking_test_subq_n0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_n3 +#### A masked pattern was here #### +POSTHOOK: query: explain +with masking_test_subq_n0 as ( select * from masking_test_n3 where key = '5') +select * from masking_test_subq_n0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_n3 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + TableScan + alias: masking_test_n3 + filterExpr: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) + Filter Operator + predicate: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) + Select Operator + expressions: 5 (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + ListSink + +PREHOOK: query: explain +with q1 as ( select * from masking_test_n3 where key = '5') +select * from masking_test_subq_n0 +PREHOOK: type: QUERY +PREHOOK: Input: default@masking_test_subq_n0 +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain +with q1 as ( select * from masking_test_n3 where key = '5') +select * from masking_test_subq_n0 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@masking_test_subq_n0 +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_subq_n0 + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: key (type: int), value (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: src + filterExpr: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: UDFToDouble(key) is not null (type: boolean) + Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), key (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string), _col1 (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + null sort order: zz + sort order: ++ + Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Semi Join 0 to 1 + keys: + 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) + 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/masking_mv.q.out ql/src/test/results/clientpositive/llap/masking_mv.q.out similarity index 55% rename from ql/src/test/results/clientpositive/masking_mv.q.out rename to ql/src/test/results/clientpositive/llap/masking_mv.q.out index 1de1a5acee..05a0f613a1 100644 --- ql/src/test/results/clientpositive/masking_mv.q.out +++ ql/src/test/results/clientpositive/llap/masking_mv.q.out @@ -28,80 +28,74 @@ POSTHOOK: Output: database:default POSTHOOK: Output: default@masking_test_view_n_mv STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-8 depends on stages: Stage-0 - Stage-2 depends on stages: Stage-8 - Stage-9 depends on stages: Stage-2 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-0, Stage-2 + Stage-3 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n_mv - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.masking_test_view_n_mv - Select Operator - expressions: _col0 (type: int) - outputColumnNames: col1 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Tez #### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n_mv + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.masking_test_view_n_mv + Select Operator + expressions: _col0 (type: int) + outputColumnNames: col1 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct) + Execution mode: llap + LLAP IO: may be used (ACID table) + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-2 + Dependency Collection - Stage: Stage-8 + Stage: Stage-4 Create View columns: key int expanded text: select `masking_test_n_mv`.`key` from `default`.`masking_test_n_mv` @@ -109,7 +103,7 @@ STAGE PLANS: original text: select key from masking_test_n_mv rewrite enabled: true - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -117,26 +111,12 @@ STAGE PLANS: Column Types: int Table: default.masking_test_view_n_mv - Stage: Stage-9 + Stage: Stage-5 Materialized View Update name: default.masking_test_view_n_mv retrieve and include: true - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-6 + Stage: Stage-0 Move Operator files: hdfs directory: true @@ -205,38 +185,22 @@ POSTHOOK: Input: default@masking_test_n_mv POSTHOOK: Input: default@masking_test_view_n_mv #### A masked pattern was here #### STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 + Stage-0 is a root stage STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: default.masking_test_view_n_mv - filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - Stage: Stage-0 Fetch Operator limit: -1 Processor Tree: - ListSink + TableScan + alias: default.masking_test_view_n_mv + filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) + Filter Operator + predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + ListSink PREHOOK: query: select key from `masking_test_n_mv` PREHOOK: type: QUERY @@ -284,41 +248,50 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: default.masking_test_view_n_mv - filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: default.masking_test_view_n_mv + filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: key (type: int) + minReductionHashAggr: 0.6 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash + keys: KEY._col0 (type: int) + mode: mergepartial outputColumnNames: _col0 Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -358,45 +331,54 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n_mv - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n_mv + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.6 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: may be used (ACID table) + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -414,9 +396,9 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n_mv POSTHOOK: Input: default@masking_test_view_n_mv_3 #### A masked pattern was here #### -0_lav 2_lav 4_lav +0_lav 8_lav PREHOOK: query: explain select value, sum(key) from `masking_test_n_mv` group by value @@ -436,48 +418,57 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n_mv - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string), key (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_n_mv + filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: reverse(value) (type: string), key (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: string) + minReductionHashAggr: 0.6 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: may be used (ACID table) + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -495,9 +486,9 @@ POSTHOOK: type: QUERY POSTHOOK: Input: default@masking_test_n_mv POSTHOOK: Input: default@masking_test_view_n_mv_3 #### A masked pattern was here #### -0_lav 0 2_lav 2 4_lav 4 +0_lav 0 8_lav 8 PREHOOK: query: create materialized view `masking_test_view_n_mv_4` as select key, value from `masking_test_n_mv` @@ -531,45 +522,54 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: default.masking_test_view_n_mv_4 - filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: default.masking_test_view_n_mv_4 + filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: reverse(value) (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + keys: _col0 (type: string) + minReductionHashAggr: 0.6 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 + Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -589,9 +589,9 @@ POSTHOOK: Input: default@masking_test_n_mv POSTHOOK: Input: default@masking_test_view_n_mv_3 POSTHOOK: Input: default@masking_test_view_n_mv_4 #### A masked pattern was here #### -0_lav 2_lav 4_lav +0_lav 8_lav PREHOOK: query: explain select value, sum(key) from `masking_test_n_mv` group by value @@ -613,48 +613,57 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: default.masking_test_view_n_mv_4 - filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string), key (type: int) + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: default.masking_test_view_n_mv_4 + filterExpr: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: ((key < 10) and (0 = (key % 2))) (type: boolean) + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: reverse(value) (type: string), key (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: sum(_col1) + keys: _col0 (type: string) + minReductionHashAggr: 0.6 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: bigint) + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Group By Operator + aggregations: sum(VALUE._col0) + keys: KEY._col0 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: sum(_col1) - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 + Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: bigint) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: sum(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 384 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator @@ -674,9 +683,9 @@ POSTHOOK: Input: default@masking_test_n_mv POSTHOOK: Input: default@masking_test_view_n_mv_3 POSTHOOK: Input: default@masking_test_view_n_mv_4 #### A masked pattern was here #### -0_lav 0 2_lav 2 4_lav 4 +0_lav 0 8_lav 8 PREHOOK: query: create table `srcTnx` stored as orc TBLPROPERTIES ('transactional'='true') as select cast(key as int) as key, value from src @@ -708,80 +717,74 @@ POSTHOOK: Output: database:default POSTHOOK: Output: default@masking_test_view_n_mv_2 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-8 depends on stages: Stage-0 - Stage-2 depends on stages: Stage-8 - Stage-9 depends on stages: Stage-2 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-0, Stage-2 + Stage-3 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-3 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srctnx - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.masking_test_view_n_mv_2 - Select Operator - expressions: _col0 (type: int) - outputColumnNames: col1 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Tez #### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srctnx + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.masking_test_view_n_mv_2 + Select Operator + expressions: _col0 (type: int) + outputColumnNames: col1 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 424 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: struct) + Execution mode: llap + LLAP IO: may be used (ACID table) + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0) + mode: mergepartial + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 440 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-8 + Stage: Stage-2 + Dependency Collection + + Stage: Stage-4 Create View columns: key int expanded text: select `srctnx`.`key` from `default`.`srcTnx` @@ -789,7 +792,7 @@ STAGE PLANS: original text: select key from srcTnx rewrite enabled: true - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -797,26 +800,12 @@ STAGE PLANS: Column Types: int Table: default.masking_test_view_n_mv_2 - Stage: Stage-9 + Stage: Stage-5 Materialized View Update name: default.masking_test_view_n_mv_2 retrieve and include: true - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-6 + Stage: Stage-0 Move Operator files: hdfs directory: true @@ -888,33 +877,42 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_view_n_mv_2 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: masking_test_view_n_mv_2 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: int) + outputColumnNames: _col0 + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + Execution mode: vectorized, llap + LLAP IO: all inputs + Reducer 2 + Execution mode: vectorized, llap + Reduce Operator Tree: + Select Operator + expressions: KEY.reducesinkkey0 (type: int) + outputColumnNames: _col0 Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Select Operator - expressions: KEY.reducesinkkey0 (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + File Output Operator + compressed: false + Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Stage: Stage-0 Fetch Operator diff --git ql/src/test/results/clientpositive/llap/masking_reserved.q.out ql/src/test/results/clientpositive/llap/masking_reserved.q.out new file mode 100644 index 0000000000..728b0f6f3b --- /dev/null +++ ql/src/test/results/clientpositive/llap/masking_reserved.q.out @@ -0,0 +1,244 @@ +PREHOOK: query: create table keyword_test_off (id int, `etad` string, key int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@keyword_test_off +POSTHOOK: query: create table keyword_test_off (id int, `etad` string, key int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@keyword_test_off +PREHOOK: query: create table keyword_test_on (id int, `date` string, key int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@keyword_test_on +POSTHOOK: query: create table keyword_test_on (id int, `date` string, key int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@keyword_test_on +PREHOOK: query: create table masking_test_n_masking_reserved (id int, value string, key int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n_masking_reserved +POSTHOOK: query: create table masking_test_n_masking_reserved (id int, value string, key int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n_masking_reserved +PREHOOK: query: create temporary table masking_test_n_masking_temp stored as orc as select * from masking_test_n_masking_reserved +PREHOOK: type: CREATETABLE_AS_SELECT +PREHOOK: Input: default@masking_test_n_masking_reserved +PREHOOK: Output: database:default +PREHOOK: Output: default@masking_test_n_masking_temp +POSTHOOK: query: create temporary table masking_test_n_masking_temp stored as orc as select * from masking_test_n_masking_reserved +POSTHOOK: type: CREATETABLE_AS_SELECT +POSTHOOK: Input: default@masking_test_n_masking_reserved +POSTHOOK: Output: database:default +POSTHOOK: Output: default@masking_test_n_masking_temp +PREHOOK: query: explain select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id +PREHOOK: type: QUERY +PREHOOK: Input: default@keyword_test_off +PREHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +POSTHOOK: query: explain select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@keyword_test_off +POSTHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), etad (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: masking_test_n_masking_reserved + filterExpr: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col1, _col3 + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id +PREHOOK: type: QUERY +PREHOOK: Input: default@keyword_test_off +PREHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +POSTHOOK: query: select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@keyword_test_off +POSTHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +PREHOOK: query: explain select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id +PREHOOK: type: QUERY +PREHOOK: Input: default@keyword_test_on +PREHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +POSTHOOK: query: explain select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@keyword_test_on +POSTHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a + filterExpr: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: id is not null (type: boolean) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), date (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 3 + Map Operator Tree: + TableScan + alias: masking_test_n_masking_reserved + filterExpr: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: id (type: int), reverse(value) (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col1, _col3 + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col1 (type: string), _col3 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id +PREHOOK: type: QUERY +PREHOOK: Input: default@keyword_test_on +PREHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### +POSTHOOK: query: select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id +POSTHOOK: type: QUERY +POSTHOOK: Input: default@keyword_test_on +POSTHOOK: Input: default@masking_test_n_masking_reserved +#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/merge3.q.out ql/src/test/results/clientpositive/llap/merge3.q.out similarity index 78% rename from ql/src/test/results/clientpositive/merge3.q.out rename to ql/src/test/results/clientpositive/llap/merge3.q.out index 32c712d9a0..e985f54235 100644 --- ql/src/test/results/clientpositive/merge3.q.out +++ ql/src/test/results/clientpositive/llap/merge3.q.out @@ -67,170 +67,164 @@ OPTIMIZED SQL: SELECT `key`, `value` FROM `default`.`merge_src` STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-8 depends on stages: Stage-0 - Stage-2 depends on stages: Stage-8 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-0, Stage-2 + Stage-3 depends on stages: Stage-4 + Stage-0 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: merge_src - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Tez #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Edges: + Reducer 2 <- Map 1 (CUSTOM_SIMPLE_EDGE) #### A masked pattern was here #### - table: + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: merge_src + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucketing_version -1 + columns key,value + columns.types string:string + name default.merge_src2 + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string) + outputColumnNames: col1, col2 + Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll') + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + null sort order: + numBuckets: -1 + sort order: + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: struct), _col1 (type: struct) + auto parallelism: false + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: merge_src + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.merge_src + numFiles 1 + numRows 2000 + rawDataSize 21248 + serialization.ddl struct merge_src { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 23248 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucketing_version -1 + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + bucketing_version 2 + column.name.delimiter , columns key,value + columns.comments columns.types string:string - name default.merge_src2 +#### A masked pattern was here #### + name default.merge_src + numFiles 1 + numRows 2000 + rawDataSize 21248 + serialization.ddl struct merge_src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: col1, col2 - Statistics: Num rows: 2000 Data size: 356000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(col1, 'hll'), compute_stats(col2, 'hll') - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - null sort order: - numBuckets: -1 - sort order: - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: struct), _col1 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: merge_src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src - numFiles 4 - numRows 2000 - rawDataSize 21248 - serialization.ddl struct merge_src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 23248 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + totalSize 23248 #### A masked pattern was here #### - name default.merge_src - numFiles 4 - numRows 2000 - rawDataSize 21248 - serialization.ddl struct merge_src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 23248 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src - name: default.merge_src - Truncated Path -> Alias: - /merge_src [merge_src] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src + name: default.merge_src + Truncated Path -> Alias: + /merge_src [merge_src] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + mode: mergepartial + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE + NumFilesPerFileSink: 1 + Statistics: Num rows: 1 Data size: 880 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1 - columns.types struct:struct - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1 + columns.types struct:struct + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false - Stage: Stage-7 - Conditional Operator + Stage: Stage-2 + Dependency Collection Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-8 Create Table columns: key string, value string name: default.merge_src2 @@ -238,7 +232,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.IgnoreKeyTextOutputFormat serde name: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -248,123 +242,7 @@ STAGE PLANS: Table: default.merge_src2 Is Table Level Stats: true - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: -1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10004 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src2 - name: default.merge_src2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: -1 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10004 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucketing_version -1 - columns key,value - columns.types string:string - name default.merge_src2 - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src2 - name: default.merge_src2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 + Stage: Stage-0 Move Operator files: hdfs directory: true @@ -2457,407 +2335,236 @@ FROM `default`.`merge_src_part` WHERE `ds` IS NOT NULL STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: merge_src_part - filterExpr: ds is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), ds (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, ds - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col1 (type: struct), _col2 (type: struct) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: ds=2008-04-08 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part - numFiles 1 - numRows 1000 - partition_columns ds - partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part - name: default.merge_src_part + Tez #### A masked pattern was here #### - Partition - base file name: ds=2008-04-09 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-09 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### - name default.merge_src_part - numFiles 1 - numRows 1000 - partition_columns ds - partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part - name: default.merge_src_part - Truncated Path -> Alias: - /merge_src_part/ds=2008-04-08 [merge_src_part] - /merge_src_part/ds=2008-04-09 [merge_src_part] - Needs Tagging: false - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: merge_src_part + filterExpr: ds is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), ds (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 1 #### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - partition: - ds - replace: true + NumFilesPerFileSink: 1 + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name default.merge_src_part2 + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - - Stage: Stage-2 - Stats Work - Basic Stats Work: + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, ds + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col0 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col1 (type: struct), _col2 (type: struct) + auto parallelism: true + Execution mode: llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.merge_src_part2 - Is Table Level Stats: false - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 + Path -> Partition: #### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: + Partition + base file name: ds=2008-04-08 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 - bucketing_version 2 column.name.delimiter , columns key,value columns.comments columns.types string:string #### A masked pattern was here #### - name default.merge_src_part2 + name default.merge_src_part + numFiles 1 + numRows 1000 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} + rawDataSize 10624 + serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11624 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - name: default.merge_src_part2 - Truncated Path -> Alias: + name default.merge_src_part + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part + name: default.merge_src_part #### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: + Partition + base file name: ds=2008-04-09 input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 - bucketing_version 2 column.name.delimiter , columns key,value columns.comments columns.types string:string #### A masked pattern was here #### - name default.merge_src_part2 + name default.merge_src_part + numFiles 1 + numRows 1000 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} + rawDataSize 10624 + serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11624 #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.merge_src_part + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - Path -> Partition: + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part + name: default.merge_src_part + Truncated Path -> Alias: + /merge_src_part/ds=2008-04-08 [merge_src_part] + /merge_src_part/ds=2008-04-09 [merge_src_part] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 #### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection + + Stage: Stage-0 + Move Operator + tables: + partition: + ds + replace: true #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - + table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: @@ -2877,15 +2584,16 @@ STAGE PLANS: #### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 - name: default.merge_src_part2 - Truncated Path -> Alias: -#### A masked pattern was here #### - Stage: Stage-6 - Move Operator - files: - hdfs directory: true + Stage: Stage-3 + Stats Work + Basic Stats Work: #### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.merge_src_part2 + Is Table Level Stats: false PREHOOK: query: insert overwrite table merge_src_part2 partition(ds) select key, value, ds from merge_src_part @@ -4963,225 +4671,224 @@ POSTHOOK: Input: default@merge_src_part@ds=2008-04-08 POSTHOOK: Input: default@merge_src_part@ds=2008-04-09 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: merge_src_part - filterExpr: ds is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Select Operator - expressions: key (type: string), value (type: string), ds (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col2 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - tag: -1 - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - auto parallelism: false - Execution mode: vectorized - Path -> Alias: + Tez #### A masked pattern was here #### - Path -> Partition: + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) #### A masked pattern was here #### - Partition - base file name: ds=2008-04-08 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part - numFiles 1 - numRows 1000 - partition_columns ds - partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: merge_src_part + filterExpr: ds is not null (type: boolean) + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + GatherStats: false + Select Operator + expressions: key (type: string), value (type: string), ds (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + bucketingVersion: 2 + key expressions: _col2 (type: string) + null sort order: z + numBuckets: -1 + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + auto parallelism: true + Execution mode: vectorized, llap + LLAP IO: no inputs + Path -> Alias: #### A masked pattern was here #### - name default.merge_src_part - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Path -> Partition: #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part - name: default.merge_src_part + Partition + base file name: ds=2008-04-08 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - Partition - base file name: ds=2008-04-09 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-09 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + name default.merge_src_part + numFiles 1 + numRows 1000 + partition_columns ds + partition_columns.types string + rawDataSize 10624 + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11624 #### A masked pattern was here #### - name default.merge_src_part - numFiles 1 - numRows 1000 - partition_columns ds - partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + name default.merge_src_part + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - name default.merge_src_part - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part + name: default.merge_src_part #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part - name: default.merge_src_part - Truncated Path -> Alias: - /merge_src_part/ds=2008-04-08 [s:merge_src_part] - /merge_src_part/ds=2008-04-09 [s:merge_src_part] - Needs Tagging: false - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 1 + Partition + base file name: ds=2008-04-09 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + name default.merge_src_part + numFiles 1 + numRows 1000 + partition_columns ds + partition_columns.types string + rawDataSize 10624 + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 11624 #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string #### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name default.merge_src_part + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, ds - Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string) - mode: complete - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part + name: default.merge_src_part + Truncated Path -> Alias: + /merge_src_part/ds=2008-04-08 [merge_src_part] + /merge_src_part/ds=2008-04-09 [merge_src_part] + Reducer 2 + Execution mode: llap + Needs Tagging: false + Reduce Operator Tree: Select Operator - expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator bucketingVersion: 2 compressed: false - GlobalTableId: 0 + GlobalTableId: 1 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucketing_version -1 - columns _col0,_col1,_col2 - columns.types struct:struct:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true + bucket_count -1 + bucketing_version 2 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.merge_src_part2 + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part2 TotalFiles: 1 - GatherStats: false + GatherStats: true MultiFileSpray: false - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + outputColumnNames: key, value, ds + Statistics: Num rows: 2000 Data size: 724000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string) + mode: complete + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col1 (type: struct), _col2 (type: struct), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + bucketingVersion: 2 + compressed: false + GlobalTableId: 0 +#### A masked pattern was here #### + NumFilesPerFileSink: 1 + Statistics: Num rows: 2 Data size: 2128 Basic stats: COMPLETE Column stats: COMPLETE #### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + bucketing_version -1 + columns _col0,_col1,_col2 + columns.types struct:struct:string + escape.delim \ + hive.serialization.extend.additional.nesting.levels true + serialization.escape.crlf true + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + TotalFiles: 1 + GatherStats: false + MultiFileSpray: false + + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -5211,7 +4918,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: #### A masked pattern was here #### @@ -5221,176 +4928,6 @@ STAGE PLANS: Table: default.merge_src_part2 Is Table Level Stats: false - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - name: default.merge_src_part2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - GatherStats: false - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: -ext-10002 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - name: default.merge_src_part2 - Truncated Path -> Alias: -#### A masked pattern was here #### - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: from (select * from merge_src_part where ds is not null distribute by ds) s insert overwrite table merge_src_part2 partition(ds) select key, value, ds diff --git ql/src/test/results/clientpositive/merge4.q.out ql/src/test/results/clientpositive/llap/merge4.q.out similarity index 81% rename from ql/src/test/results/clientpositive/merge4.q.out rename to ql/src/test/results/clientpositive/llap/merge4.q.out index 8b9740105d..e15a91c038 100644 --- ql/src/test/results/clientpositive/merge4.q.out +++ ql/src/test/results/clientpositive/llap/merge4.q.out @@ -21,79 +21,79 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 456000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 2316 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -108,7 +108,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -116,36 +116,6 @@ STAGE PLANS: Column Types: string, string Table: default.nzhang_part - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table nzhang_part partition (ds='2010-08-15', hr) select key, value, hr from srcpart where ds='2008-04-08' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart @@ -1191,79 +1161,79 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=11 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -1278,7 +1248,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -1286,36 +1256,6 @@ STAGE PLANS: Column Types: string, string Table: default.nzhang_part - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table nzhang_part partition (ds='2010-08-15', hr=11) select key, value from srcpart where ds='2008-04-08' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart @@ -2867,98 +2807,123 @@ POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0, Stage-8 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 - Stage-8 depends on stages: Stage-1 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 1500 Data size: 491500 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '1' (type: string), '1' (type: string), 'file,' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 129500 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 1500 Data size: 491500 Basic stats: COMPLETE Column stats: COMPLETE + Tez +#### A masked pattern was here #### + Edges: + Map 1 <- Union 2 (CONTAINS) + Map 5 <- Union 2 (CONTAINS) + Reducer 3 <- Union 2 (CUSTOM_SIMPLE_EDGE) + Reducer 4 <- Reducer 3 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 362000 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 5 + Map Operator Tree: + TableScan + alias: src + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: '1' (type: string), '1' (type: string), 'file,' (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 129500 Basic stats: COMPLETE Column stats: COMPLETE + Limit + Number of rows: 1 + Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + null sort order: + sort order: + Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE + TopN Hash Memory Usage: 0.1 + value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE Limit Number of rows: 1 Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: + File Output Operator + compressed: false Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 1 - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 362 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2010-08-15' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1 Data size: 456 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.0 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-7 - Conditional Operator + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Union 2 + Vertex: Union 2 - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -2973,7 +2938,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -2981,67 +2946,6 @@ STAGE PLANS: Column Types: string, string Table: default.nzhang_part - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-8 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1158 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - PREHOOK: query: insert overwrite table nzhang_part partition (ds='2010-08-15', hr) select * from ( select key, value, hr from srcpart where ds='2008-04-08' @@ -3063,9 +2967,9 @@ POSTHOOK: Input: default@src POSTHOOK: Input: default@srcpart POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=file, -POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=file,).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=file,).value EXPRESSION [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] +POSTHOOK: Output: default@nzhang_part@ds=2010-08-15/hr=11 +POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=11).key EXPRESSION [(srcpart)srcpart.FieldSchema(name:key, type:string, comment:default), ] +POSTHOOK: Lineage: nzhang_part PARTITION(ds=2010-08-15,hr=11).value EXPRESSION [(srcpart)srcpart.FieldSchema(name:value, type:string, comment:default), ] PREHOOK: query: show partitions nzhang_part PREHOOK: type: SHOWPARTITIONS PREHOOK: Input: default@nzhang_part @@ -3074,15 +2978,11 @@ POSTHOOK: type: SHOWPARTITIONS POSTHOOK: Input: default@nzhang_part ds=2010-08-15/hr=11 ds=2010-08-15/hr=12 -ds=2010-08-15/hr=file, PREHOOK: query: select * from nzhang_part where hr = 'file,' PREHOOK: type: QUERY PREHOOK: Input: default@nzhang_part -PREHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=file, #### A masked pattern was here #### POSTHOOK: query: select * from nzhang_part where hr = 'file,' POSTHOOK: type: QUERY POSTHOOK: Input: default@nzhang_part -POSTHOOK: Input: default@nzhang_part@ds=2010-08-15/hr=file, #### A masked pattern was here #### -1 1 2010-08-15 file, diff --git ql/src/test/results/clientpositive/merge_dynamic_partition.q.out ql/src/test/results/clientpositive/llap/merge_dynamic_partition.q.out similarity index 77% rename from ql/src/test/results/clientpositive/merge_dynamic_partition.q.out rename to ql/src/test/results/clientpositive/llap/merge_dynamic_partition.q.out index 4439ffa54f..72db3e800c 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition.q.out +++ ql/src/test/results/clientpositive/llap/merge_dynamic_partition.q.out @@ -66,83 +66,88 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_n1 - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_n1 + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col2 (type: struct), _col3 (type: struct) + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col2 (type: struct), _col3 (type: struct) - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part_n1 Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: a - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - Dp Sort State: PARTITION_SORTED - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 + Dependency Collection Stage: Stage-0 Move Operator @@ -718,79 +723,79 @@ POSTHOOK: Input: default@srcpart_merge_dp_n1@ds=2008-04-08/hr=11 POSTHOOK: Output: default@merge_dynamic_part_n1@ds=2008-04-08/hr=11 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_n1 - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_n1 + filterExpr: (ds = '2008-04-08') (type: boolean) Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator + File Output Operator + compressed: false + Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part_n1 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), '11' (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 99 Data size: 31648 Basic stats: PARTIAL Column stats: NONE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 49 Data size: 15664 Basic stats: PARTIAL Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -805,7 +810,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_dynamic_part_n1 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -813,36 +818,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part_n1 - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part_n1 partition (ds='2008-04-08', hr=11) select key, value from srcpart_merge_dp_n1 where ds='2008-04-08' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart_merge_dp_n1 @@ -1396,102 +1371,93 @@ POSTHOOK: Input: default@srcpart_merge_dp_n1@ds=2008-04-08/hr=11 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_n1 - filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D)) (type: boolean) - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_n1 + filterExpr: ((ds = '2008-04-08') and (11.0D = 11.0D)) (type: boolean) + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: key (type: string), value (type: string), '2008-04-08' (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col2 (type: struct), _col3 (type: struct) + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: aa + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 99 Data size: 136984 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col2 (type: struct), _col3 (type: struct) - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 49 Data size: 65384 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part_n1 Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: string), _col3 (type: string) - null sort order: aa - sort order: ++ - Map-reduce partition columns: _col2 (type: string), _col3 (type: string) - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - Dp Sort State: PARTITION_SORTED - Statistics: Num rows: 99 Data size: 49864 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Dependency Collection Stage: Stage-0 Move Operator @@ -1514,36 +1480,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part_n1 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n1 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part_n1 partition (ds, hr) select key, value, ds, hr from srcpart_merge_dp_n1 where ds='2008-04-08' and hr=11 PREHOOK: type: QUERY PREHOOK: Input: default@srcpart_merge_dp_n1 diff --git ql/src/test/results/clientpositive/merge_dynamic_partition2.q.out ql/src/test/results/clientpositive/llap/merge_dynamic_partition2.q.out similarity index 57% rename from ql/src/test/results/clientpositive/merge_dynamic_partition2.q.out rename to ql/src/test/results/clientpositive/llap/merge_dynamic_partition2.q.out index 8d189dd158..9bf582529b 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition2.q.out +++ ql/src/test/results/clientpositive/llap/merge_dynamic_partition2.q.out @@ -80,102 +80,93 @@ POSTHOOK: Input: default@srcpart_merge_dp_n0@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_n0 - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: key (type: string), value (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_n0 + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: key (type: string), value (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 297 Data size: 409848 Basic stats: PARTIAL Column stats: PARTIAL + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 297 Data size: 409848 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col2 (type: struct), _col3 (type: struct) + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 297 Data size: 409848 Basic stats: PARTIAL Column stats: PARTIAL - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 297 Data size: 409848 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col2 (type: struct), _col3 (type: struct) - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 148 Data size: 196480 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part_n0 Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: string) - null sort order: a - sort order: + - Map-reduce partition columns: _col2 (type: string) - Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - File Output Operator - compressed: false - Dp Sort State: PARTITION_SORTED - Statistics: Num rows: 297 Data size: 148488 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n0 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Dependency Collection Stage: Stage-0 Move Operator @@ -198,36 +189,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part_n0 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n0 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n0 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part_n0 partition (ds='2008-04-08', hr) select key, value, hr from srcpart_merge_dp_n0 where ds='2008-04-08' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart_merge_dp_n0 diff --git ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out ql/src/test/results/clientpositive/llap/merge_dynamic_partition3.q.out similarity index 69% rename from ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out rename to ql/src/test/results/clientpositive/llap/merge_dynamic_partition3.q.out index c46d0e8b78..d8b4c0dbd2 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out +++ ql/src/test/results/clientpositive/llap/merge_dynamic_partition3.q.out @@ -144,102 +144,93 @@ POSTHOOK: Input: default@srcpart_merge_dp_n2@ds=2008-04-09/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage Stage-2 depends on stages: Stage-1 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_n2 - filterExpr: (ds >= '2008-04-08') (type: boolean) - Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_n2 + filterExpr: (ds >= '2008-04-08') (type: boolean) + Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 594 Data size: 928256 Basic stats: PARTIAL Column stats: PARTIAL + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 594 Data size: 928256 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col2 (type: struct), _col3 (type: struct) + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: aa + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) + Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL + value expressions: _col0 (type: string), _col1 (type: string) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 594 Data size: 928256 Basic stats: PARTIAL Column stats: PARTIAL - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 594 Data size: 928256 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col2 (type: struct), _col3 (type: struct) - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL - File Output Operator - compressed: false - Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL + File Output Operator + compressed: false + Statistics: Num rows: 297 Data size: 448304 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part_n2 Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col2 (type: string), _col3 (type: string) - null sort order: aa - sort order: ++ - Map-reduce partition columns: _col2 (type: string), _col3 (type: string) - Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL - value expressions: _col0 (type: string), _col1 (type: string) - Reduce Operator Tree: - Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - File Output Operator - compressed: false - Dp Sort State: PARTITION_SORTED - Statistics: Num rows: 594 Data size: 405536 Basic stats: PARTIAL Column stats: PARTIAL - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n2 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Dependency Collection Stage: Stage-0 Move Operator @@ -262,36 +253,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part_n2 - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n2 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part_n2 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part_n2 partition (ds, hr) select key, value, ds, hr from srcpart_merge_dp_n2 where ds>='2008-04-08' PREHOOK: type: QUERY PREHOOK: Input: default@srcpart_merge_dp_n2 diff --git ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out ql/src/test/results/clientpositive/llap/merge_dynamic_partition4.q.out similarity index 73% rename from ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out rename to ql/src/test/results/clientpositive/llap/merge_dynamic_partition4.q.out index 8387880d6b..49b44ac297 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out +++ ql/src/test/results/clientpositive/llap/merge_dynamic_partition4.q.out @@ -140,79 +140,79 @@ POSTHOOK: Input: default@srcpart_merge_dp_rc_n1@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart_merge_dp_rc_n1@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_rc_n1 - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 2.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.merge_dynamic_part_n3 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_rc_n1 + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 2.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1000 Data size: 264000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.merge_dynamic_part_n3 + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 1000 Data size: 358000 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -227,7 +227,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.merge_dynamic_part_n3 - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -235,26 +235,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part_n3 - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part_n3 partition (ds = '2008-04-08', hr) select key, value, if(key % 2 == 0, 'a1', 'b1') as hr from srcpart_merge_dp_rc_n1 where ds = '2008-04-08' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out ql/src/test/results/clientpositive/llap/merge_dynamic_partition5.q.out similarity index 70% rename from ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out rename to ql/src/test/results/clientpositive/llap/merge_dynamic_partition5.q.out index 518f400ba1..1b2567894f 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out +++ ql/src/test/results/clientpositive/llap/merge_dynamic_partition5.q.out @@ -116,79 +116,79 @@ POSTHOOK: Input: default@srcpart_merge_dp_rc@ds=2008-04-08/hr=11 POSTHOOK: Input: default@srcpart_merge_dp_rc@ds=2008-04-08/hr=12 STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-2 depends on stages: Stage-1 + Stage-0 depends on stages: Stage-2 + Stage-3 depends on stages: Stage-0 STAGE PLANS: Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: srcpart_merge_dp_rc - filterExpr: (ds = '2008-04-08') (type: boolean) - Statistics: Num rows: 618 Data size: 110004 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 618 Data size: 163152 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 618 Data size: 163152 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.merge_dynamic_part - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 618 Data size: 221244 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: srcpart_merge_dp_rc + filterExpr: (ds = '2008-04-08') (type: boolean) + Statistics: Num rows: 618 Data size: 110004 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0D) = 0.0D), 'a1', 'b1') (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 618 Data size: 163152 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 618 Data size: 163152 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.merge_dynamic_part + Select Operator + expressions: _col0 (type: string), _col1 (type: string), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: key, value, ds, hr + Statistics: Num rows: 618 Data size: 221244 Basic stats: COMPLETE Column stats: COMPLETE + Group By Operator + aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') + keys: ds (type: string), hr (type: string) + minReductionHashAggr: 0.99 + mode: hash + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string), _col1 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col2 (type: struct), _col3 (type: struct) + Execution mode: llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Group By Operator + aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) + keys: KEY._col0 (type: string), KEY._col1 (type: string) + mode: mergepartial + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) + File Output Operator + compressed: false Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Stage: Stage-2 + Dependency Collection Stage: Stage-0 Move Operator @@ -203,7 +203,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.merge_dynamic_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: Column Stats Desc: @@ -211,26 +211,6 @@ STAGE PLANS: Column Types: string, string Table: default.merge_dynamic_part - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - RCFile Merge Operator - merge level: block - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table merge_dynamic_part partition (ds = '2008-04-08', hr) select key, value, if(key % 100 == 0, 'a1', 'b1') as hr from srcpart_merge_dp_rc where ds = '2008-04-08' PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/merge_empty.q.out ql/src/test/results/clientpositive/llap/merge_empty.q.out similarity index 97% rename from ql/src/test/results/clientpositive/merge_empty.q.out rename to ql/src/test/results/clientpositive/llap/merge_empty.q.out index 5d851dc181..37f6057691 100644 --- ql/src/test/results/clientpositive/merge_empty.q.out +++ ql/src/test/results/clientpositive/llap/merge_empty.q.out @@ -41,5 +41,3 @@ POSTHOOK: Input: default@bar_n1 POSTHOOK: Output: default@foo_n6 POSTHOOK: Lineage: foo_n6.a SIMPLE [(bar_n1)bar_n1.FieldSchema(name:a, type:bigint, comment:null), ] POSTHOOK: Lineage: foo_n6.b SIMPLE [(bar_n1)bar_n1.FieldSchema(name:b, type:string, comment:null), ] -Found 256 items -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/llap/merge_join_1.q.out ql/src/test/results/clientpositive/llap/merge_join_1.q.out new file mode 100644 index 0000000000..49ac5c2897 --- /dev/null +++ ql/src/test/results/clientpositive/llap/merge_join_1.q.out @@ -0,0 +1,206 @@ +PREHOOK: query: drop table if exists test_join_1 +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table if exists test_join_1 +POSTHOOK: type: DROPTABLE +PREHOOK: query: drop table if exists test_join_2 +PREHOOK: type: DROPTABLE +POSTHOOK: query: drop table if exists test_join_2 +POSTHOOK: type: DROPTABLE +PREHOOK: query: create table test_join_1(a string, b string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_join_1 +POSTHOOK: query: create table test_join_1(a string, b string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@test_join_1 +PREHOOK: query: create table test_join_2(a string, b string) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@test_join_2 +POSTHOOK: query: create table test_join_2(a string, b string) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@test_join_2 +PREHOOK: query: explain +select * from +( + SELECT a a, b b + FROM test_join_1 +)t1 + +join + +( + SELECT a a, b b + FROM test_join_1 +)t2 + on t1.a = t2.a + and t1.a = t2.b + +join + +( + select a from test_join_2 +)t3 on t1.a = t3.a +PREHOOK: type: QUERY +PREHOOK: Input: default@test_join_1 +PREHOOK: Input: default@test_join_2 +#### A masked pattern was here #### +POSTHOOK: query: explain +select * from +( + SELECT a a, b b + FROM test_join_1 +)t1 + +join + +( + SELECT a a, b b + FROM test_join_1 +)t2 + on t1.a = t2.a + and t1.a = t2.b + +join + +( + select a from test_join_2 +)t3 on t1.a = t3.a +POSTHOOK: type: QUERY +POSTHOOK: Input: default@test_join_1 +POSTHOOK: Input: default@test_join_2 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: test_join_1 + filterExpr: (a is not null or (a is not null and b is not null)) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: a is not null (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string) + Filter Operator + predicate: (a is not null and b is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string), b (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: string), _col0 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col1 (type: string), _col0 (type: string) + Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: test_join_2 + filterExpr: a is not null (type: boolean) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: a is not null (type: boolean) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: a (type: string) + outputColumnNames: _col0 + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: string), _col0 (type: string) + null sort order: zz + sort order: ++ + Map-reduce partition columns: _col0 (type: string), _col0 (type: string) + Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: string), _col2 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string), _col0 (type: string) + 1 _col1 (type: string), _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: drop table test_join_1 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@test_join_1 +PREHOOK: Output: default@test_join_1 +POSTHOOK: query: drop table test_join_1 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@test_join_1 +POSTHOOK: Output: default@test_join_1 +PREHOOK: query: drop table test_join_2 +PREHOOK: type: DROPTABLE +PREHOOK: Input: default@test_join_2 +PREHOOK: Output: default@test_join_2 +POSTHOOK: query: drop table test_join_2 +POSTHOOK: type: DROPTABLE +POSTHOOK: Input: default@test_join_2 +POSTHOOK: Output: default@test_join_2 diff --git ql/src/test/results/clientpositive/llap/mergejoins.q.out ql/src/test/results/clientpositive/llap/mergejoins.q.out new file mode 100644 index 0000000000..f08188d04a --- /dev/null +++ ql/src/test/results/clientpositive/llap/mergejoins.q.out @@ -0,0 +1,386 @@ +PREHOOK: query: create table a_n0 (val1 int, val2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@a_n0 +POSTHOOK: query: create table a_n0 (val1 int, val2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@a_n0 +PREHOOK: query: create table b_n0 (val1 int, val2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@b_n0 +POSTHOOK: query: create table b_n0 (val1 int, val2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@b_n0 +PREHOOK: query: create table c_n0 (val1 int, val2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@c_n0 +POSTHOOK: query: create table c_n0 (val1 int, val2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@c_n0 +PREHOOK: query: create table d_n0 (val1 int, val2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@d_n0 +POSTHOOK: query: create table d_n0 (val1 int, val2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@d_n0 +PREHOOK: query: create table e_n0 (val1 int, val2 int) +PREHOOK: type: CREATETABLE +PREHOOK: Output: database:default +PREHOOK: Output: default@e_n0 +POSTHOOK: query: create table e_n0 (val1 int, val2 int) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: database:default +POSTHOOK: Output: default@e_n0 +PREHOOK: query: explain select * from a_n0 join b_n0 on a_n0.val1=b_n0.val1 join c_n0 on a_n0.val1=c_n0.val1 join d_n0 on a_n0.val1=d_n0.val1 join e_n0 on a_n0.val2=e_n0.val2 +PREHOOK: type: QUERY +PREHOOK: Input: default@a_n0 +PREHOOK: Input: default@b_n0 +PREHOOK: Input: default@c_n0 +PREHOOK: Input: default@d_n0 +PREHOOK: Input: default@e_n0 +#### A masked pattern was here #### +POSTHOOK: query: explain select * from a_n0 join b_n0 on a_n0.val1=b_n0.val1 join c_n0 on a_n0.val1=c_n0.val1 join d_n0 on a_n0.val1=d_n0.val1 join e_n0 on a_n0.val2=e_n0.val2 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@a_n0 +POSTHOOK: Input: default@b_n0 +POSTHOOK: Input: default@c_n0 +POSTHOOK: Input: default@d_n0 +POSTHOOK: Input: default@e_n0 +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 6 (SIMPLE_EDGE) + Reducer 3 <- Map 7 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) + Reducer 4 <- Map 8 (SIMPLE_EDGE), Reducer 3 (SIMPLE_EDGE) + Reducer 5 <- Map 9 (SIMPLE_EDGE), Reducer 4 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: a_n0 + filterExpr: (val1 is not null and val2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: (val1 is not null and val2 is not null) (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val1 (type: int), val2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 6 + Map Operator Tree: + TableScan + alias: b_n0 + filterExpr: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val1 (type: int), val2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 7 + Map Operator Tree: + TableScan + alias: c_n0 + filterExpr: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val1 (type: int), val2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 8 + Map Operator Tree: + TableScan + alias: d_n0 + filterExpr: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val1 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val1 (type: int), val2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 9 + Map Operator Tree: + TableScan + alias: e_n0 + filterExpr: val2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Filter Operator + predicate: val2 is not null (type: boolean) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: val1 (type: int), val2 (type: int) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col0 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int) + Reducer 4 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: int) + 1 _col0 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col1 (type: int) + null sort order: z + sort order: + + Map-reduce partition columns: _col1 (type: int) + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int) + Reducer 5 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col1 (type: int) + 1 _col1 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + +PREHOOK: query: explain select * from src a_n0 join src b_n0 on a_n0.key=b_n0.key left outer join src c_n0 on b_n0.key=c_n0.key and b_n0.key<10 +PREHOOK: type: QUERY +PREHOOK: Input: default@src +#### A masked pattern was here #### +POSTHOOK: query: explain select * from src a_n0 join src b_n0 on a_n0.key=b_n0.key left outer join src c_n0 on b_n0.key=c_n0.key and b_n0.key<10 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@src +#### A masked pattern was here #### +STAGE DEPENDENCIES: + Stage-1 is a root stage + Stage-0 depends on stages: Stage-1 + +STAGE PLANS: + Stage: Stage-1 + Tez +#### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE), Map 4 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE), Reducer 2 (SIMPLE_EDGE) +#### A masked pattern was here #### + Vertices: + Map 1 + Map Operator Tree: + TableScan + alias: b_n0 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string), (UDFToDouble(key) < 10.0D) (type: boolean) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean) + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Map 4 + Map Operator Tree: + TableScan + alias: a_n0 + filterExpr: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Filter Operator + predicate: key is not null (type: boolean) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: key (type: string), value (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string) + Execution mode: vectorized, llap + LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Inner Join 0 to 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4 + Statistics: Num rows: 791 Data size: 284760 Basic stats: COMPLETE Column stats: COMPLETE + Reduce Output Operator + key expressions: _col0 (type: string) + null sort order: z + sort order: + + Map-reduce partition columns: _col0 (type: string) + Statistics: Num rows: 791 Data size: 284760 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col1 (type: string), _col2 (type: boolean), _col3 (type: string), _col4 (type: string) + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Merge Join Operator + condition map: + Left Outer Join 0 to 1 + filter predicates: + 0 {VALUE._col1} + 1 + keys: + 0 _col0 (type: string) + 1 _col0 (type: string) + outputColumnNames: _col0, _col1, _col3, _col4, _col5, _col6 + Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE + Select Operator + expressions: _col3 (type: string), _col4 (type: string), _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 + Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + Stage: Stage-0 + Fetch Operator + limit: -1 + Processor Tree: + ListSink + diff --git ql/src/test/results/clientpositive/llap_reader.q.out ql/src/test/results/clientpositive/llap_reader.q.out deleted file mode 100644 index ec07da97bf..0000000000 --- ql/src/test/results/clientpositive/llap_reader.q.out +++ /dev/null @@ -1,88 +0,0 @@ -PREHOOK: query: CREATE TABLE test_n7(f1 int, f2 int, f3 int) stored as orc -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_n7 -PREHOOK: query: INSERT INTO TABLE test_n7 VALUES (1,1,1), (2,2,2), (3,3,3), (4,4,4) -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@test_n7 -PREHOOK: query: ALTER TABLE test_n7 CHANGE f1 f1 bigint -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_n7 -PREHOOK: Output: default@test_n7 -PREHOOK: query: ALTER TABLE test_n7 CHANGE f2 f2 bigint -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_n7 -PREHOOK: Output: default@test_n7 -PREHOOK: query: ALTER TABLE test_n7 CHANGE f3 f3 bigint -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_n7 -PREHOOK: Output: default@test_n7 -PREHOOK: query: SELECT count(f1) FROM test_n7 GROUP BY f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_n7 -#### A masked pattern was here #### -1 -1 -1 -1 -PREHOOK: query: SELECT count(f1) FROM test_n7 GROUP BY f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_n7 -#### A masked pattern was here #### -1 -1 -1 -1 -PREHOOK: query: CREATE TABLE test_bigint(f1 bigint, f2 bigint, f3 bigint) stored as orc -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_bigint -PREHOOK: query: INSERT OVERWRITE TABLE test_bigint select * from test_n7 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_n7 -PREHOOK: Output: default@test_bigint -PREHOOK: query: ALTER TABLE test_bigint CHANGE f1 f1 double -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_bigint -PREHOOK: Output: default@test_bigint -PREHOOK: query: ALTER TABLE test_bigint CHANGE f2 f2 double -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_bigint -PREHOOK: Output: default@test_bigint -PREHOOK: query: ALTER TABLE test_bigint CHANGE f3 f3 double -PREHOOK: type: ALTERTABLE_RENAMECOL -PREHOOK: Input: default@test_bigint -PREHOOK: Output: default@test_bigint -PREHOOK: query: select count(f1) from test_bigint group by f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_bigint -#### A masked pattern was here #### -1 -1 -1 -1 -PREHOOK: query: select count(f1) from test_bigint group by f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_bigint -#### A masked pattern was here #### -1 -1 -1 -1 -PREHOOK: query: CREATE TABLE test_acid_n0 (f1 int, f2 int, val string) clustered by (val) into 2 buckets stored as orc TBLPROPERTIES ('transactional'='true') -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_acid_n0 -PREHOOK: query: INSERT INTO TABLE test_acid_n0 VALUES (1,1,'b1'), (2,2,'b2'), (3,3,'b3'), (4,4,'b4') -PREHOOK: type: QUERY -PREHOOK: Input: _dummy_database@_dummy_table -PREHOOK: Output: default@test_acid_n0 -PREHOOK: query: SELECT count(f1) FROM test_acid_n0 GROUP BY f1 -PREHOOK: type: QUERY -PREHOOK: Input: default@test_acid_n0 -#### A masked pattern was here #### -1 -1 -1 -1 diff --git ql/src/test/results/clientpositive/load_dyn_part13.q.out ql/src/test/results/clientpositive/load_dyn_part13.q.out deleted file mode 100644 index 917bedce22..0000000000 --- ql/src/test/results/clientpositive/load_dyn_part13.q.out +++ /dev/null @@ -1,270 +0,0 @@ -PREHOOK: query: show partitions srcpart -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@srcpart -POSTHOOK: query: show partitions srcpart -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@srcpart -ds=2008-04-08/hr=11 -ds=2008-04-08/hr=12 -ds=2008-04-09/hr=11 -ds=2008-04-09/hr=12 -PREHOOK: query: create table if not exists nzhang_part13 like srcpart -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_part13 -POSTHOOK: query: create table if not exists nzhang_part13 like srcpart -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_part13 -PREHOOK: query: describe extended nzhang_part13 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_part13 -POSTHOOK: query: describe extended nzhang_part13 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_part13 -key string default -value string default -ds string -hr string - -# Partition Information -# col_name data_type comment -ds string -hr string - -#### A masked pattern was here #### -PREHOOK: query: explain -insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) -select * from ( - select key, value, '22' - from src - where key < 20 - union all - select key, value, '33' - from src - where key > 20 and key < 40) s -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@nzhang_part13@ds=2010-03-03 -POSTHOOK: query: explain -insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) -select * from ( - select key, value, '22' - from src - where key < 20 - union all - select key, value, '33' - from src - where key > 20 and key < 40) s -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - Stage-2 depends on stages: Stage-0 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (UDFToDouble(key) < 20.0D) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) < 20.0D) (type: boolean) - Statistics: Num rows: 166 Data size: 29548 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '22' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 166 Data size: 43824 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part13 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2010-03-03' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 221 Data size: 79118 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - TableScan - alias: src - filterExpr: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 40.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((UDFToDouble(key) > 20.0D) and (UDFToDouble(key) < 40.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), '33' (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 55 Data size: 14520 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 221 Data size: 58344 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part13 - Select Operator - expressions: _col0 (type: string), _col1 (type: string), '2010-03-03' (type: string), _col2 (type: string) - outputColumnNames: key, value, ds, hr - Statistics: Num rows: 221 Data size: 79118 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col2 (type: struct), _col3 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 1060 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Move Operator - tables: - partition: - ds 2010-03-03 - hr - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part13 - - Stage: Stage-2 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.nzhang_part13 - -PREHOOK: query: insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) -select * from ( - select key, value, '22' - from src - where key < 20 - union all - select key, value, '33' - from src - where key > 20 and key < 40) s -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@nzhang_part13@ds=2010-03-03 -POSTHOOK: query: insert overwrite table nzhang_part13 partition (ds="2010-03-03", hr) -select * from ( - select key, value, '22' - from src - where key < 20 - union all - select key, value, '33' - from src - where key > 20 and key < 40) s -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@nzhang_part13@ds=2010-03-03/hr=22 -POSTHOOK: Output: default@nzhang_part13@ds=2010-03-03/hr=33 -POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=22).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: nzhang_part13 PARTITION(ds=2010-03-03,hr=33).value EXPRESSION [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: show partitions nzhang_part13 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@nzhang_part13 -POSTHOOK: query: show partitions nzhang_part13 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@nzhang_part13 -ds=2010-03-03/hr=22 -ds=2010-03-03/hr=33 -PREHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_part13 -PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22 -PREHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33 -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_part13 where ds is not null and hr is not null -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_part13 -POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=22 -POSTHOOK: Input: default@nzhang_part13@ds=2010-03-03/hr=33 -#### A masked pattern was here #### -15 val_15 2010-03-03 22 -17 val_17 2010-03-03 22 -0 val_0 2010-03-03 22 -4 val_4 2010-03-03 22 -12 val_12 2010-03-03 22 -8 val_8 2010-03-03 22 -0 val_0 2010-03-03 22 -0 val_0 2010-03-03 22 -15 val_15 2010-03-03 22 -19 val_19 2010-03-03 22 -10 val_10 2010-03-03 22 -5 val_5 2010-03-03 22 -11 val_11 2010-03-03 22 -5 val_5 2010-03-03 22 -2 val_2 2010-03-03 22 -12 val_12 2010-03-03 22 -5 val_5 2010-03-03 22 -18 val_18 2010-03-03 22 -9 val_9 2010-03-03 22 -18 val_18 2010-03-03 22 -27 val_27 2010-03-03 33 -37 val_37 2010-03-03 33 -35 val_35 2010-03-03 33 -24 val_24 2010-03-03 33 -26 val_26 2010-03-03 33 -35 val_35 2010-03-03 33 -34 val_34 2010-03-03 33 -30 val_30 2010-03-03 33 -33 val_33 2010-03-03 33 -35 val_35 2010-03-03 33 -24 val_24 2010-03-03 33 -26 val_26 2010-03-03 33 -28 val_28 2010-03-03 33 -37 val_37 2010-03-03 33 diff --git ql/src/test/results/clientpositive/load_dyn_part14.q.out ql/src/test/results/clientpositive/load_dyn_part14.q.out deleted file mode 100644 index 24ddee5daf..0000000000 --- ql/src/test/results/clientpositive/load_dyn_part14.q.out +++ /dev/null @@ -1,376 +0,0 @@ -PREHOOK: query: create table if not exists nzhang_part14_n0 (key string) - partitioned by (value string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@nzhang_part14_n0 -POSTHOOK: query: create table if not exists nzhang_part14_n0 (key string) - partitioned by (value string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@nzhang_part14_n0 -PREHOOK: query: describe extended nzhang_part14_n0 -PREHOOK: type: DESCTABLE -PREHOOK: Input: default@nzhang_part14_n0 -POSTHOOK: query: describe extended nzhang_part14_n0 -POSTHOOK: type: DESCTABLE -POSTHOOK: Input: default@nzhang_part14_n0 -key string -value string - -# Partition Information -# col_name data_type comment -value string - -#### A masked pattern was here #### -PREHOOK: query: explain -insert overwrite table nzhang_part14_n0 partition(value) -select key, value from ( - select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a - union all - select * from (select 'k2' as key, '' as value from src limit 2)b - union all - select * from (select 'k3' as key, ' ' as value from src limit 2)c -) T -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@nzhang_part14_n0 -POSTHOOK: query: explain -insert overwrite table nzhang_part14_n0 partition(value) -select key, value from ( - select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a - union all - select * from (select 'k2' as key, '' as value from src limit 2)b - union all - select * from (select 'k3' as key, ' ' as value from src limit 2)c -) T -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-9, Stage-10 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 - Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-9 is a root stage - Stage-10 is a root stage - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'k1' (type: string), null (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 256 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Union - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll') - keys: value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct) - TableScan - Union - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll') - keys: value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct) - TableScan - Union - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - Select Operator - expressions: _col0 (type: string), _col1 (type: string) - outputColumnNames: key, value - Statistics: Num rows: 6 Data size: 771 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - aggregations: compute_stats(key, 'hll') - keys: value (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: struct) - Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0) - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: struct), _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 1050 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-0 - Move Operator - tables: - partition: - value - replace: true - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - - Stage: Stage-3 - Stats Work - Basic Stats Work: - Column Stats Desc: - Columns: key - Column Types: string - Table: default.nzhang_part14_n0 - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14_n0 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'k2' (type: string), '' (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - alias: src - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 2 Data size: 8 Basic stats: COMPLETE Column stats: COMPLETE - TopN Hash Memory Usage: 0.1 - Execution mode: vectorized - Reduce Operator Tree: - Limit - Number of rows: 2 - Statistics: Num rows: 2 Data size: 16 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 'k3' (type: string), ' ' (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - -PREHOOK: query: insert overwrite table nzhang_part14_n0 partition(value) -select key, value from ( - select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a - union all - select * from (select 'k2' as key, '' as value from src limit 2)b - union all - select * from (select 'k3' as key, ' ' as value from src limit 2)c -) T -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Output: default@nzhang_part14_n0 -POSTHOOK: query: insert overwrite table nzhang_part14_n0 partition(value) -select key, value from ( - select * from (select 'k1' as key, cast(null as string) as value from src limit 2)a - union all - select * from (select 'k2' as key, '' as value from src limit 2)b - union all - select * from (select 'k3' as key, ' ' as value from src limit 2)c -) T -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Output: default@nzhang_part14_n0@value= -POSTHOOK: Output: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ -POSTHOOK: Lineage: nzhang_part14_n0 PARTITION(value= ).key EXPRESSION [] -POSTHOOK: Lineage: nzhang_part14_n0 PARTITION(value=__HIVE_DEFAULT_PARTITION__).key EXPRESSION [] -PREHOOK: query: show partitions nzhang_part14_n0 -PREHOOK: type: SHOWPARTITIONS -PREHOOK: Input: default@nzhang_part14_n0 -POSTHOOK: query: show partitions nzhang_part14_n0 -POSTHOOK: type: SHOWPARTITIONS -POSTHOOK: Input: default@nzhang_part14_n0 -value= -value=__HIVE_DEFAULT_PARTITION__ -PREHOOK: query: select * from nzhang_part14_n0 where value <> 'a' -PREHOOK: type: QUERY -PREHOOK: Input: default@nzhang_part14_n0 -PREHOOK: Input: default@nzhang_part14_n0@value= -PREHOOK: Input: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ -#### A masked pattern was here #### -POSTHOOK: query: select * from nzhang_part14_n0 where value <> 'a' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@nzhang_part14_n0 -POSTHOOK: Input: default@nzhang_part14_n0@value= -POSTHOOK: Input: default@nzhang_part14_n0@value=__HIVE_DEFAULT_PARTITION__ -#### A masked pattern was here #### -k1 __HIVE_DEFAULT_PARTITION__ -k1 __HIVE_DEFAULT_PARTITION__ -k2 __HIVE_DEFAULT_PARTITION__ -k2 __HIVE_DEFAULT_PARTITION__ -k3 -k3 diff --git ql/src/test/results/clientpositive/louter_join_ppr.q.out ql/src/test/results/clientpositive/louter_join_ppr.q.out deleted file mode 100644 index a116abe193..0000000000 --- ql/src/test/results/clientpositive/louter_join_ppr.q.out +++ /dev/null @@ -1,1292 +0,0 @@ -PREHOOK: query: EXPLAIN EXTENDED - FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key AND b.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED - FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key AND b.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` < 20 AND `key` > 15) AS `t0` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `key` > 15 AND `key` < 20 AND `ds` = '2008-04-08') AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: string) - auto parallelism: false - TableScan - alias: b - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_0:a] - /srcpart/ds=2008-04-08/hr=11 [$hdt$_1:b] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key AND b.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key AND b.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -17 val_17 17 val_17 -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -19 val_19 19 val_19 -PREHOOK: query: EXPLAIN EXTENDED - FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key AND a.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED - FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key AND a.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `key` < 20 AND `key` > 15 AND `ds` = '2008-04-08') AS `t0` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` > 15 AND `key` < 20) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: string) - auto parallelism: false - TableScan - alias: b - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:b] - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:a] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key AND a.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key AND a.ds = '2008-04-08') - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -17 val_17 17 val_17 -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -19 val_19 19 val_19 -PREHOOK: query: EXPLAIN EXTENDED - FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED - FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` < 20 AND `key` > 15) AS `t0` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `key` > 15 AND `key` < 20 AND `ds` = '2008-04-08') AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: string) - auto parallelism: false - TableScan - alias: b - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_0:a] - /srcpart/ds=2008-04-08/hr=11 [$hdt$_1:b] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_1:b] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: FROM - src a - LEFT OUTER JOIN - srcpart b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND b.ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -17 val_17 17 val_17 -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -19 val_19 19 val_19 -PREHOOK: query: EXPLAIN EXTENDED - FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: EXPLAIN EXTENDED - FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -OPTIMIZED SQL: SELECT * -FROM (SELECT `key`, `value` -FROM `default`.`srcpart` -WHERE `key` < 20 AND `key` > 15 AND `ds` = '2008-04-08') AS `t0` -INNER JOIN (SELECT `key`, `value` -FROM `default`.`src` -WHERE `key` > 15 AND `key` < 20) AS `t2` ON `t0`.`key` = `t2`.`key` -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 1000 Data size: 178000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) < 20.0D) and (UDFToDouble(key) > 15.0D)) (type: boolean) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 111 Data size: 19758 Basic stats: COMPLETE Column stats: COMPLETE - tag: 0 - value expressions: _col1 (type: string) - auto parallelism: false - TableScan - alias: b - filterExpr: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - GatherStats: false - Filter Operator - isSamplingPred: false - predicate: ((UDFToDouble(key) > 15.0D) and (UDFToDouble(key) < 20.0D)) (type: boolean) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - bucketingVersion: 2 - key expressions: _col0 (type: string) - null sort order: z - numBuckets: -1 - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 55 Data size: 9790 Basic stats: COMPLETE Column stats: COMPLETE - tag: 1 - value expressions: _col1 (type: string) - auto parallelism: false - Path -> Alias: -#### A masked pattern was here #### - Path -> Partition: -#### A masked pattern was here #### - Partition - base file name: src - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.src - name: default.src -#### A masked pattern was here #### - Partition - base file name: hr=11 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 11 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart -#### A masked pattern was here #### - Partition - base file name: hr=12 - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - partition values: - ds 2008-04-08 - hr 12 - properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - numFiles 1 - numRows 500 - partition_columns ds/hr - partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - bucketing_version 2 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.srcpart - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.srcpart - name: default.srcpart - Truncated Path -> Alias: - /src [$hdt$_1:b] - /srcpart/ds=2008-04-08/hr=11 [$hdt$_0:a] - /srcpart/ds=2008-04-08/hr=12 [$hdt$_0:a] - Needs Tagging: true - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - bucketingVersion: 2 - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 55 Data size: 19580 Basic stats: COMPLETE Column stats: COMPLETE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - bucketing_version -1 - columns _col0,_col1,_col2,_col3 - columns.types string:string:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' -PREHOOK: type: QUERY -PREHOOK: Input: default@src -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: FROM - srcpart a - LEFT OUTER JOIN - src b - ON (a.key = b.key) - SELECT a.key, a.value, b.key, b.value - WHERE a.key > 10 AND a.key < 20 AND b.key > 15 AND b.key < 25 AND a.ds = '2008-04-08' -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -#### A masked pattern was here #### -17 val_17 17 val_17 -17 val_17 17 val_17 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -18 val_18 18 val_18 -19 val_19 19 val_19 -19 val_19 19 val_19 diff --git ql/src/test/results/clientpositive/masking_1.q.out ql/src/test/results/clientpositive/masking_1.q.out deleted file mode 100644 index 9a72b41b59..0000000000 --- ql/src/test/results/clientpositive/masking_1.q.out +++ /dev/null @@ -1,530 +0,0 @@ -PREHOOK: query: create table masking_test_n8 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n8 -POSTHOOK: query: create table masking_test_n8 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n8 -POSTHOOK: Lineage: masking_test_n8.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n8.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select * from masking_test_n8 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n8 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n8 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n8 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -0 0_lav -4 4_lav -8 8_lav -0 0_lav -0 0_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select key from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select key from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select key from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -4 -8 -2 -PREHOOK: query: explain select value from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select value from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select value from masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select value from masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -4_lav -8_lav -2_lav -PREHOOK: query: explain select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and ((key % 2) = 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5 Data size: 980 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col2 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col2 (type: double) - Statistics: Num rows: 5 Data size: 980 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - alias: srcpart - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), ds (type: string), hr (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col4 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col4 (type: double) - Statistics: Num rows: 2000 Data size: 1108000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col2 (type: double) - 1 _col4 (type: double) - outputColumnNames: _col0, _col1, _col3, _col4, _col5, _col6 - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n8 join srcpart on (masking_test_n8.key = srcpart.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -2 2_lav 2 val_2 2008-04-09 11 -2 2_lav 2 val_2 2008-04-08 11 -2 2_lav 2 val_2 2008-04-09 12 -2 2_lav 2 val_2 2008-04-08 12 -4 4_lav 4 val_4 2008-04-08 12 -4 4_lav 4 val_4 2008-04-09 12 -4 4_lav 4 val_4 2008-04-08 11 -4 4_lav 4 val_4 2008-04-09 11 -8 8_lav 8 val_8 2008-04-08 11 -8 8_lav 8 val_8 2008-04-09 11 -8 8_lav 8 val_8 2008-04-08 12 -8 8_lav 8 val_8 2008-04-09 12 -PREHOOK: query: explain select * from default.masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from default.masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from default.masking_test_n8 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select * from default.masking_test_n8 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n8 where masking_test_n8.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n8 where masking_test_n8.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n8 - filterExpr: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 0) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n8 where masking_test_n8.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n8 where masking_test_n8.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n8 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToDouble(key) % 2.0D) = 0.0D) and (UDFToDouble(key) > 0.0D) and (UDFToDouble(key) < 10.0D)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), upper(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_10.q.out ql/src/test/results/clientpositive/masking_10.q.out deleted file mode 100644 index 5c4db51b85..0000000000 --- ql/src/test/results/clientpositive/masking_10.q.out +++ /dev/null @@ -1,266 +0,0 @@ -PREHOOK: query: drop table masking_test -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table masking_test -POSTHOOK: type: DROPTABLE -PREHOOK: query: create temporary table masking_test as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test -POSTHOOK: query: create temporary table masking_test as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test -PREHOOK: query: explain select 2017 as a, value from masking_test group by 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test -#### A masked pattern was here #### -POSTHOOK: query: explain select 2017 as a, value from masking_test group by 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: reverse(value) (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: 2017 (type: int), _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select 2017 as a, value from masking_test group by 1, 2 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test -#### A masked pattern was here #### -POSTHOOK: query: select 2017 as a, value from masking_test group by 1, 2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test -#### A masked pattern was here #### -2017 0_lav -2017 2_lav -2017 4_lav -2017 8_lav -Warning: Shuffle Join JOIN[37][tables = [$hdt$_0, $hdt$_1]] in Stage 'Stage-1:MAPRED' is a cross product -PREHOOK: query: explain -select * from - masking_test alias01 - left join - ( - select 2017 as a, value from masking_test group by 1, 2 - ) alias02 - on alias01.key = alias02.a - left join - masking_test alias03 -on alias01.key = alias03.key -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from - masking_test alias01 - left join - ( - select 2017 as a, value from masking_test group by 1, 2 - ) alias02 - on alias01.key = alias02.a - left join - masking_test alias03 -on alias01.key = alias03.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-1 depends on stages: Stage-3 - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: reverse(value) (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Group By Operator - keys: _col0 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: string) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 125 Data size: 22372 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 0 - Statistics: Num rows: 0 Data size: 0 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: 2017 (type: int), _col0 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - Reduce Output Operator - null sort order: - sort order: - Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE - value expressions: _col0 (type: int), _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 - 1 - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 0 Data size: 0 Basic stats: NONE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: int), _col3 (type: string) - TableScan - alias: masking_test - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89488 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 44744 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 275 Data size: 49218 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 275 Data size: 49218 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_12.q.out ql/src/test/results/clientpositive/masking_12.q.out deleted file mode 100644 index ad0ed643bb..0000000000 --- ql/src/test/results/clientpositive/masking_12.q.out +++ /dev/null @@ -1,501 +0,0 @@ -PREHOOK: query: create table `masking_test_n5` as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n5 -POSTHOOK: query: create table `masking_test_n5` as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n5 -POSTHOOK: Lineage: masking_test_n5.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n5.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create view `v0` as select * from `masking_test_n5` -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Output: database:default -PREHOOK: Output: default@v0 -POSTHOOK: query: create view `v0` as select * from `masking_test_n5` -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v0 -POSTHOOK: Lineage: v0.key SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:key, type:int, comment:null), ] -POSTHOOK: Lineage: v0.value SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:value, type:string, comment:null), ] -PREHOOK: query: explain -select * from `v0` -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from `v0` -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n5 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from `v0` -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: select * from `v0` -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -0 0_lav -4 4_lav -8 8_lav -0 0_lav -0 0_lav -2 2_lav -PREHOOK: query: create table `masking_test_subq_n1` as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_subq_n1 -POSTHOOK: query: create table `masking_test_subq_n1` as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_subq_n1 -POSTHOOK: Lineage: masking_test_subq_n1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_subq_n1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create view `v1_n9` as select * from `masking_test_subq_n1` -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@masking_test_subq_n1 -PREHOOK: Output: database:default -PREHOOK: Output: default@v1_n9 -POSTHOOK: query: create view `v1_n9` as select * from `masking_test_subq_n1` -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@masking_test_subq_n1 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@v1_n9 -POSTHOOK: Lineage: v1_n9.key SIMPLE [(masking_test_subq_n1)masking_test_subq_n1.FieldSchema(name:key, type:int, comment:null), ] -POSTHOOK: Lineage: v1_n9.value SIMPLE [(masking_test_subq_n1)masking_test_subq_n1.FieldSchema(name:value, type:string, comment:null), ] -PREHOOK: query: explain -select * from `v1_n9` -limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n1 -PREHOOK: Input: default@src -PREHOOK: Input: default@v1_n9 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from `v1_n9` -limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n1 -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1_n9 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n1 - filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n1 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - Limit - Number of rows: 20 - Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 20 Data size: 1900 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: 20 - Processor Tree: - ListSink - -PREHOOK: query: select * from `v1_n9` -limit 20 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n1 -PREHOOK: Input: default@src -PREHOOK: Input: default@v1_n9 -#### A masked pattern was here #### -POSTHOOK: query: select * from `v1_n9` -limit 20 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n1 -POSTHOOK: Input: default@src -POSTHOOK: Input: default@v1_n9 -#### A masked pattern was here #### -0 val_0 -0 val_0 -0 val_0 -2 val_2 -4 val_4 -5 val_5 -5 val_5 -5 val_5 -8 val_8 -9 val_9 -10 val_10 -11 val_11 -12 val_12 -12 val_12 -15 val_15 -15 val_15 -17 val_17 -18 val_18 -18 val_18 -19 val_19 -PREHOOK: query: create view `masking_test_view` as select key from `v0` -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@v0 -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_view -POSTHOOK: query: create view `masking_test_view` as select key from `v0` -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@v0 -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_view -POSTHOOK: Lineage: masking_test_view.key SIMPLE [(masking_test_n5)masking_test_n5.FieldSchema(name:key, type:int, comment:null), ] -PREHOOK: query: explain -select key from `masking_test_view` -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@masking_test_view -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: explain -select key from `masking_test_view` -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@masking_test_view -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n5 - filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 6) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key from `masking_test_view` -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@masking_test_view -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: select key from `masking_test_view` -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@masking_test_view -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -4 -PREHOOK: query: explain -select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@masking_test_view -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: explain -select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@masking_test_view -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n5 - filterExpr: ((key < 10) and ((key % 2) = 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: masking_test_n5 - filterExpr: ((key < 10) and (key > 6) and ((key % 2) = 0) and ((UDFToInteger((UDFToDouble(key) / 2.0D)) % 2) = 0) and (UDFToInteger((UDFToDouble(key) / 2.0D)) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key < 10) and (key > 6) and ((key % 2) = 0) and ((UDFToInteger((UDFToDouble(key) / 2.0D)) % 2) = 0) and (UDFToInteger((UDFToDouble(key) / 2.0D)) < 10)) (type: boolean) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger((UDFToDouble(key) / 2.0D)) (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 4 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col1 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col1 (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n5 -PREHOOK: Input: default@masking_test_view -PREHOOK: Input: default@v0 -#### A masked pattern was here #### -POSTHOOK: query: select `v0`.value from `v0` join `masking_test_view` on `v0`.key = `masking_test_view`.key -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n5 -POSTHOOK: Input: default@masking_test_view -POSTHOOK: Input: default@v0 -#### A masked pattern was here #### -4_lav diff --git ql/src/test/results/clientpositive/masking_2.q.out ql/src/test/results/clientpositive/masking_2.q.out deleted file mode 100644 index 550a0fd97f..0000000000 --- ql/src/test/results/clientpositive/masking_2.q.out +++ /dev/null @@ -1,385 +0,0 @@ -PREHOOK: query: create view masking_test_n1 as select cast(key as int) as key, value from src -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n1 -POSTHOOK: query: create view masking_test_n1 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n1 -POSTHOOK: Lineage: masking_test_n1.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n1.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select * from masking_test_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n1 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 0_lav -4 4_lav -8 8_lav -0 0_lav -0 0_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n1 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n1 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n1 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n1 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select * from src a join masking_test_n1 b on a.key = b.value where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from src a join masking_test_n1 b on a.key = b.value where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and reverse(value) is not null) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and reverse(value) is not null) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select a.*, b.key from masking_test_n1 a join masking_test_n1 b on a.key = b.value where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select a.*, b.key from masking_test_n1 a join masking_test_n1 b on a.key = b.value where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col0) (type: double) - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and UDFToDouble(reverse(value)) is not null) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0) and UDFToDouble(reverse(value)) is not null) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col1) (type: double) - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double) - 1 UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * from masking_test_n1 a union select b.* from masking_test_n1 b where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n1 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n1 a union select b.* from masking_test_n1 b where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n1 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 110 Data size: 20680 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: int), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 110 Data size: 20680 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: int), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_4.q.out ql/src/test/results/clientpositive/masking_4.q.out deleted file mode 100644 index 8a73950198..0000000000 --- ql/src/test/results/clientpositive/masking_4.q.out +++ /dev/null @@ -1,331 +0,0 @@ -PREHOOK: query: create table masking_test_n11 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n11 -POSTHOOK: query: create table masking_test_n11 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n11 -POSTHOOK: Lineage: masking_test_n11.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n11.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create table masking_test_subq_n2 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_subq_n2 -POSTHOOK: query: create table masking_test_subq_n2 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_subq_n2 -POSTHOOK: Lineage: masking_test_subq_n2.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_subq_n2.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '5') (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '5') (type: boolean) - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '5' (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 170 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with q1 as ( select * from masking_test_n11 where key = '5') -select * from q1 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n11 -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select * from masking_test_n11 where key = '5') -select * from q1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n11 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n11 - filterExpr: ((key = 5) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key = 5) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 1 Data size: 95 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 5 (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with masking_test_subq_n2 as ( select * from masking_test_n11 where key = '5') -select * from masking_test_subq_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n11 -#### A masked pattern was here #### -POSTHOOK: query: explain -with masking_test_subq_n2 as ( select * from masking_test_n11 where key = '5') -select * from masking_test_subq_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n11 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n11 - filterExpr: ((key = 5) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key = 5) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 1 Data size: 95 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 5 (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with q1 as ( select * from masking_test_n11 where key = '5') -select * from masking_test_subq_n2 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n2 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select * from masking_test_n11 where key = '5') -select * from masking_test_subq_n2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n2 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-3 is a root stage - Stage-2 depends on stages: Stage-3 - Stage-1 depends on stages: Stage-2 - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n2 - filterExpr: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (UDFToDouble(key) is not null and key is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: key (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Execution mode: vectorized - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int) - mode: mergepartial - outputColumnNames: _col0 - Statistics: Num rows: 250 Data size: 1000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), UDFToDouble(_col0) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: string) - TableScan - Reduce Output Operator - key expressions: _col1 (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: double) - Statistics: Num rows: 250 Data size: 3000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: double) - 1 _col1 (type: double) - outputColumnNames: _col0, _col2 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 395 Data size: 35945 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: int) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n2 - filterExpr: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key is not null and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col0 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col0 (type: int) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - TableScan - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), _col1 (type: int) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), _col1 (type: int) - Statistics: Num rows: 197 Data size: 17927 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double), _col0 (type: int) - 1 UDFToDouble(_col0) (type: double), _col1 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 325 Data size: 30875 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_7.q.out ql/src/test/results/clientpositive/masking_7.q.out deleted file mode 100644 index 89984dcaef..0000000000 --- ql/src/test/results/clientpositive/masking_7.q.out +++ /dev/null @@ -1,295 +0,0 @@ -PREHOOK: query: drop view masking_test_n14 -PREHOOK: type: DROPVIEW -POSTHOOK: query: drop view masking_test_n14 -POSTHOOK: type: DROPVIEW -PREHOOK: query: create view masking_test_n14 as select cast(key as int) as key, value, '12', ROW__ID from src -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n14 -POSTHOOK: query: create view masking_test_n14 as select cast(key as int) as key, value, '12', ROW__ID from src -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n14 -POSTHOOK: Lineage: masking_test_n14._c2 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n14.row__id SIMPLE [(src)src.FieldSchema(name:ROW__ID, type:struct, comment:), ] -POSTHOOK: Lineage: masking_test_n14.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select * from masking_test_n14 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n14 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string), ROW__ID (type: struct) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 83 Data size: 29050 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 29050 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n14 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n14 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 0_lav 12 NULL -4 4_lav 12 NULL -8 8_lav 12 NULL -0 0_lav 12 NULL -0 0_lav 12 NULL -2 2_lav 12 NULL -PREHOOK: query: explain select * from masking_test_n14 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n14 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string), '12' (type: string), ROW__ID (type: struct) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 27 Data size: 9450 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 9450 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n14 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n14 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -4 4_lav 12 NULL -8 8_lav 12 NULL -2 2_lav 12 NULL -PREHOOK: query: drop view masking_test_n14 -PREHOOK: type: DROPVIEW -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Output: default@masking_test_n14 -POSTHOOK: query: drop view masking_test_n14 -POSTHOOK: type: DROPVIEW -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Output: default@masking_test_n14 -PREHOOK: query: create view masking_test_n14 as select cast(key as int) as key, '12', ROW__ID, -'12', '12', '12', '12', '12', '12', '12', '12', '12', '12' - from src -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n14 -POSTHOOK: query: create view masking_test_n14 as select cast(key as int) as key, '12', ROW__ID, -'12', '12', '12', '12', '12', '12', '12', '12', '12', '12' - from src -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n14 -POSTHOOK: Lineage: masking_test_n14._c1 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c10 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c11 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c12 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c3 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c4 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c5 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c6 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c7 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c8 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14._c9 SIMPLE [] -POSTHOOK: Lineage: masking_test_n14.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n14.row__id SIMPLE [(src)src.FieldSchema(name:ROW__ID, type:struct, comment:), ] -PREHOOK: query: explain select * from masking_test_n14 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n14 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 7221 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), '12' (type: string), ROW__ID (type: struct), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 83 Data size: 85158 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 85158 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n14 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n14 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 12 NULL 12 12 12 12 12 12 12 12 12 12 -4 12 NULL 12 12 12 12 12 12 12 12 12 12 -8 12 NULL 12 12 12 12 12 12 12 12 12 12 -0 12 NULL 12 12 12 12 12 12 12 12 12 12 -0 12 NULL 12 12 12 12 12 12 12 12 12 12 -2 12 NULL 12 12 12 12 12 12 12 12 12 12 -PREHOOK: query: explain select * from masking_test_n14 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n14 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - properties: - insideView TRUE - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 2349 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), '12' (type: string), ROW__ID (type: struct), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string), '12' (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9, _col10, _col11, _col12 - Statistics: Num rows: 27 Data size: 27702 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 27702 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n14 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n14 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n14 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n14 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -4 12 NULL 12 12 12 12 12 12 12 12 12 12 -8 12 NULL 12 12 12 12 12 12 12 12 12 12 -2 12 NULL 12 12 12 12 12 12 12 12 12 12 diff --git ql/src/test/results/clientpositive/masking_disablecbo_1.q.out ql/src/test/results/clientpositive/masking_disablecbo_1.q.out deleted file mode 100644 index c030374ebe..0000000000 --- ql/src/test/results/clientpositive/masking_disablecbo_1.q.out +++ /dev/null @@ -1,526 +0,0 @@ -PREHOOK: query: create table masking_test_n10 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n10 -POSTHOOK: query: create table masking_test_n10 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n10 -POSTHOOK: Lineage: masking_test_n10.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n10.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select * from masking_test_n10 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n10 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -0 0_lav -4 4_lav -8 8_lav -0 0_lav -0 0_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select key from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select key from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 2000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 20 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select key from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select key from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -4 -8 -2 -PREHOOK: query: explain select value from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select value from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: reverse(value) (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 5 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 920 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select value from masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select value from masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -4_lav -8_lav -2_lav -PREHOOK: query: explain select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and UDFToDouble(key) is not null) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 2280 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col0) (type: double) - Statistics: Num rows: 5 Data size: 2280 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - alias: srcpart - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(key) (type: double) - Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double) - 1 UDFToDouble(key) (type: double) - outputColumnNames: _col0, _col1, _col5, _col6, _col7, _col8 - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: int), _col1 (type: string), _col5 (type: string), _col6 (type: string), _col7 (type: string), _col8 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 31 Data size: 22754 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -PREHOOK: Input: default@srcpart -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -PREHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n10 join srcpart on (masking_test_n10.key = srcpart.key) -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -POSTHOOK: Input: default@srcpart -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-08/hr=12 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=11 -POSTHOOK: Input: default@srcpart@ds=2008-04-09/hr=12 -#### A masked pattern was here #### -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-08 11 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-08 12 -0 0_lav 0 val_0 2008-04-09 11 -0 0_lav 0 val_0 2008-04-09 12 -2 2_lav 2 val_2 2008-04-09 11 -2 2_lav 2 val_2 2008-04-08 11 -2 2_lav 2 val_2 2008-04-09 12 -2 2_lav 2 val_2 2008-04-08 12 -4 4_lav 4 val_4 2008-04-08 12 -4 4_lav 4 val_4 2008-04-09 12 -4 4_lav 4 val_4 2008-04-08 11 -4 4_lav 4 val_4 2008-04-09 11 -8 8_lav 8 val_8 2008-04-08 11 -8 8_lav 8 val_8 2008-04-09 11 -8 8_lav 8 val_8 2008-04-08 12 -8 8_lav 8 val_8 2008-04-09 12 -PREHOOK: query: explain select * from default.masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from default.masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from default.masking_test_n10 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select * from default.masking_test_n10 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n10 where masking_test_n10.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n10 where masking_test_n10.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n10 - filterExpr: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key > 0)) (type: boolean) - Statistics: Num rows: 5 Data size: 475 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 5 Data size: 940 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n10 where masking_test_n10.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n10 where masking_test_n10.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n10 -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select key, value from (select key, value from (select key, upper(value) as value from src where key > 0) t where key < 10) t2 where key % 2 = 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: ((key > 0) and (key < 10) and ((key % 2) = 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), upper(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 7317 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_disablecbo_2.q.out ql/src/test/results/clientpositive/masking_disablecbo_2.q.out deleted file mode 100644 index 4ae035b582..0000000000 --- ql/src/test/results/clientpositive/masking_disablecbo_2.q.out +++ /dev/null @@ -1,371 +0,0 @@ -PREHOOK: query: create view masking_test_n13 as select cast(key as int) as key, value from src -PREHOOK: type: CREATEVIEW -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n13 -POSTHOOK: query: create view masking_test_n13 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATEVIEW -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n13 -POSTHOOK: Lineage: masking_test_n13.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n13.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain select * from masking_test_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n13 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n13 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -0 0_lav -4 4_lav -8 8_lav -0 0_lav -0 0_lav -2 2_lav -PREHOOK: query: explain select * from masking_test_n13 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n13 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select * from masking_test_n13 where key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: select * from masking_test_n13 where key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -4 4_lav -8 8_lav -2 2_lav -PREHOOK: query: explain select * from src a join masking_test_n13 b on a.key = b.value where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from src a join masking_test_n13 b on a.key = b.value where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and reverse(value) is not null and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and reverse(value) is not null and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col1 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: string) - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - TableScan - alias: a - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: key (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: key (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: value (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 key (type: string) - 1 _col1 (type: string) - outputColumnNames: _col0, _col1, _col5, _col6 - Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col5 (type: int), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 42 Data size: 15372 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select a.*, b.key from masking_test_n13 a join masking_test_n13 b on a.key = b.value where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select a.*, b.key from masking_test_n13 a join masking_test_n13 b on a.key = b.value where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(UDFToInteger(key)) is not null) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col0) (type: double) - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int), _col1 (type: string) - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(reverse(value)) is not null and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and UDFToDouble(reverse(value)) is not null and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col1) (type: double) - null sort order: z - sort order: + - Map-reduce partition columns: UDFToDouble(_col1) (type: double) - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 UDFToDouble(_col0) (type: double) - 1 UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 27 Data size: 5184 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * from masking_test_n13 a union select b.* from masking_test_n13 b where b.key > 0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n13 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from masking_test_n13 a union select b.* from masking_test_n13 b where b.key > 0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n13 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10)) (type: boolean) - Statistics: Num rows: 83 Data size: 14774 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 83 Data size: 15604 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 110 Data size: 20680 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: int), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - TableScan - alias: src - filterExpr: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((UDFToInteger(key) % 2) = 0) and (UDFToInteger(key) < 10) and (UDFToInteger(key) > 0)) (type: boolean) - Statistics: Num rows: 27 Data size: 4806 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: UDFToInteger(key) (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 27 Data size: 5076 Basic stats: COMPLETE Column stats: COMPLETE - Union - Statistics: Num rows: 110 Data size: 20680 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: int), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: int), _col1 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: int), _col1 (type: string) - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Group By Operator - keys: KEY._col0 (type: int), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1 - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 55 Data size: 10340 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_disablecbo_4.q.out ql/src/test/results/clientpositive/masking_disablecbo_4.q.out deleted file mode 100644 index 584d0df8b0..0000000000 --- ql/src/test/results/clientpositive/masking_disablecbo_4.q.out +++ /dev/null @@ -1,246 +0,0 @@ -PREHOOK: query: create table masking_test_n3 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n3 -POSTHOOK: query: create table masking_test_n3 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n3 -POSTHOOK: Lineage: masking_test_n3.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_n3.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: create table masking_test_subq_n0 as select cast(key as int) as key, value from src -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@src -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_subq_n0 -POSTHOOK: query: create table masking_test_subq_n0 as select cast(key as int) as key, value from src -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@src -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_subq_n0 -POSTHOOK: Lineage: masking_test_subq_n0.key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] -POSTHOOK: Lineage: masking_test_subq_n0.value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] -PREHOOK: query: explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select key from q2 where key = '5'), -q2 as ( select key from src where key = '5') -select * from (select key from q1) a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: src - filterExpr: (key = '5') (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (key = '5') (type: boolean) - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: '5' (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 174 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with q1 as ( select * from masking_test_n3 where key = '5') -select * from q1 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n3 -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select * from masking_test_n3 where key = '5') -select * from q1 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n3 - filterExpr: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) - Statistics: Num rows: 1 Data size: 95 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 5 (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with masking_test_subq_n0 as ( select * from masking_test_n3 where key = '5') -select * from masking_test_subq_n0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_n3 -#### A masked pattern was here #### -POSTHOOK: query: explain -with masking_test_subq_n0 as ( select * from masking_test_n3 where key = '5') -select * from masking_test_subq_n0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_n3 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_n3 - filterExpr: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: (((key % 2) = 0) and (key < 10) and (key = 5)) (type: boolean) - Statistics: Num rows: 1 Data size: 95 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: 5 (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - Execution mode: vectorized - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain -with q1 as ( select * from masking_test_n3 where key = '5') -select * from masking_test_subq_n0 -PREHOOK: type: QUERY -PREHOOK: Input: default@masking_test_subq_n0 -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain -with q1 as ( select * from masking_test_n3 where key = '5') -select * from masking_test_subq_n0 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@masking_test_subq_n0 -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: masking_test_subq_n0 - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - Statistics: Num rows: 500 Data size: 47500 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: key (type: int), value (type: string) - TableScan - alias: src - filterExpr: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: UDFToDouble(key) is not null (type: boolean) - Statistics: Num rows: 500 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), key (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE - Group By Operator - keys: _col0 (type: string), _col1 (type: string) - minReductionHashAggr: 0.99 - mode: hash - outputColumnNames: _col0, _col1 - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - null sort order: zz - sort order: ++ - Map-reduce partition columns: UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - Statistics: Num rows: 250 Data size: 43500 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Operator Tree: - Join Operator - condition map: - Left Semi Join 0 to 1 - keys: - 0 UDFToDouble(key) (type: double), UDFToDouble(key) (type: double) - 1 UDFToDouble(_col0) (type: double), UDFToDouble(_col1) (type: double) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 412 Data size: 39140 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - diff --git ql/src/test/results/clientpositive/masking_reserved.q.out ql/src/test/results/clientpositive/masking_reserved.q.out deleted file mode 100644 index 11d81cde39..0000000000 --- ql/src/test/results/clientpositive/masking_reserved.q.out +++ /dev/null @@ -1,216 +0,0 @@ -PREHOOK: query: create table keyword_test_off (id int, `etad` string, key int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@keyword_test_off -POSTHOOK: query: create table keyword_test_off (id int, `etad` string, key int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@keyword_test_off -PREHOOK: query: create table keyword_test_on (id int, `date` string, key int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@keyword_test_on -POSTHOOK: query: create table keyword_test_on (id int, `date` string, key int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@keyword_test_on -PREHOOK: query: create table masking_test_n_masking_reserved (id int, value string, key int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n_masking_reserved -POSTHOOK: query: create table masking_test_n_masking_reserved (id int, value string, key int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n_masking_reserved -PREHOOK: query: create temporary table masking_test_n_masking_temp stored as orc as select * from masking_test_n_masking_reserved -PREHOOK: type: CREATETABLE_AS_SELECT -PREHOOK: Input: default@masking_test_n_masking_reserved -PREHOOK: Output: database:default -PREHOOK: Output: default@masking_test_n_masking_temp -POSTHOOK: query: create temporary table masking_test_n_masking_temp stored as orc as select * from masking_test_n_masking_reserved -POSTHOOK: type: CREATETABLE_AS_SELECT -POSTHOOK: Input: default@masking_test_n_masking_reserved -POSTHOOK: Output: database:default -POSTHOOK: Output: default@masking_test_n_masking_temp -PREHOOK: query: explain select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id -PREHOOK: type: QUERY -PREHOOK: Input: default@keyword_test_off -PREHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -POSTHOOK: query: explain select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@keyword_test_off -POSTHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), etad (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: masking_test_n_masking_reserved - filterExpr: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id -PREHOOK: type: QUERY -PREHOOK: Input: default@keyword_test_off -PREHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -POSTHOOK: query: select a.`etad`, b.value from keyword_test_off a join masking_test_n_masking_reserved b on b.id = a.id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@keyword_test_off -POSTHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -PREHOOK: query: explain select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id -PREHOOK: type: QUERY -PREHOOK: Input: default@keyword_test_on -PREHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -POSTHOOK: query: explain select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@keyword_test_on -POSTHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-0 depends on stages: Stage-1 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a - filterExpr: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: id is not null (type: boolean) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), date (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 188 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: masking_test_n_masking_reserved - filterExpr: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: ((key < 10) and ((key % 2) = 0) and id is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: id (type: int), reverse(value) (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 192 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col1, _col3 - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col1 (type: string), _col3 (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 206 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id -PREHOOK: type: QUERY -PREHOOK: Input: default@keyword_test_on -PREHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### -POSTHOOK: query: select a.`date`, b.value from keyword_test_on a join masking_test_n_masking_reserved b on b.id = a.id -POSTHOOK: type: QUERY -POSTHOOK: Input: default@keyword_test_on -POSTHOOK: Input: default@masking_test_n_masking_reserved -#### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/merge_join_1.q.out ql/src/test/results/clientpositive/merge_join_1.q.out deleted file mode 100644 index 6703703744..0000000000 --- ql/src/test/results/clientpositive/merge_join_1.q.out +++ /dev/null @@ -1,205 +0,0 @@ -PREHOOK: query: drop table if exists test_join_1 -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table if exists test_join_1 -POSTHOOK: type: DROPTABLE -PREHOOK: query: drop table if exists test_join_2 -PREHOOK: type: DROPTABLE -POSTHOOK: query: drop table if exists test_join_2 -POSTHOOK: type: DROPTABLE -PREHOOK: query: create table test_join_1(a string, b string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_join_1 -POSTHOOK: query: create table test_join_1(a string, b string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@test_join_1 -PREHOOK: query: create table test_join_2(a string, b string) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@test_join_2 -POSTHOOK: query: create table test_join_2(a string, b string) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@test_join_2 -PREHOOK: query: explain -select * from -( - SELECT a a, b b - FROM test_join_1 -)t1 - -join - -( - SELECT a a, b b - FROM test_join_1 -)t2 - on t1.a = t2.a - and t1.a = t2.b - -join - -( - select a from test_join_2 -)t3 on t1.a = t3.a -PREHOOK: type: QUERY -PREHOOK: Input: default@test_join_1 -PREHOOK: Input: default@test_join_2 -#### A masked pattern was here #### -POSTHOOK: query: explain -select * from -( - SELECT a a, b b - FROM test_join_1 -)t1 - -join - -( - SELECT a a, b b - FROM test_join_1 -)t2 - on t1.a = t2.a - and t1.a = t2.b - -join - -( - select a from test_join_2 -)t3 on t1.a = t3.a -POSTHOOK: type: QUERY -POSTHOOK: Input: default@test_join_1 -POSTHOOK: Input: default@test_join_2 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: test_join_1 - filterExpr: a is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: a is not null (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) - TableScan - alias: test_join_2 - filterExpr: a is not null (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: a is not null (type: boolean) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string) - outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 1 Data size: 184 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string), _col0 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col0 (type: string) - Statistics: Num rows: 1 Data size: 404 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string), _col2 (type: string) - TableScan - alias: test_join_1 - filterExpr: (a is not null and b is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (a is not null and b is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: a (type: string), b (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: string), _col0 (type: string) - null sort order: zz - sort order: ++ - Map-reduce partition columns: _col1 (type: string), _col0 (type: string) - Statistics: Num rows: 1 Data size: 368 Basic stats: COMPLETE Column stats: NONE - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string), _col0 (type: string) - 1 _col1 (type: string), _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col3 (type: string), _col4 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 444 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: drop table test_join_1 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@test_join_1 -PREHOOK: Output: default@test_join_1 -POSTHOOK: query: drop table test_join_1 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@test_join_1 -POSTHOOK: Output: default@test_join_1 -PREHOOK: query: drop table test_join_2 -PREHOOK: type: DROPTABLE -PREHOOK: Input: default@test_join_2 -PREHOOK: Output: default@test_join_2 -POSTHOOK: query: drop table test_join_2 -POSTHOOK: type: DROPTABLE -POSTHOOK: Input: default@test_join_2 -POSTHOOK: Output: default@test_join_2 diff --git ql/src/test/results/clientpositive/mergejoins.q.out ql/src/test/results/clientpositive/mergejoins.q.out deleted file mode 100644 index 4283e1a167..0000000000 --- ql/src/test/results/clientpositive/mergejoins.q.out +++ /dev/null @@ -1,389 +0,0 @@ -PREHOOK: query: create table a_n0 (val1 int, val2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@a_n0 -POSTHOOK: query: create table a_n0 (val1 int, val2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@a_n0 -PREHOOK: query: create table b_n0 (val1 int, val2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@b_n0 -POSTHOOK: query: create table b_n0 (val1 int, val2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@b_n0 -PREHOOK: query: create table c_n0 (val1 int, val2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@c_n0 -POSTHOOK: query: create table c_n0 (val1 int, val2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@c_n0 -PREHOOK: query: create table d_n0 (val1 int, val2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@d_n0 -POSTHOOK: query: create table d_n0 (val1 int, val2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@d_n0 -PREHOOK: query: create table e_n0 (val1 int, val2 int) -PREHOOK: type: CREATETABLE -PREHOOK: Output: database:default -PREHOOK: Output: default@e_n0 -POSTHOOK: query: create table e_n0 (val1 int, val2 int) -POSTHOOK: type: CREATETABLE -POSTHOOK: Output: database:default -POSTHOOK: Output: default@e_n0 -PREHOOK: query: explain select * from a_n0 join b_n0 on a_n0.val1=b_n0.val1 join c_n0 on a_n0.val1=c_n0.val1 join d_n0 on a_n0.val1=d_n0.val1 join e_n0 on a_n0.val2=e_n0.val2 -PREHOOK: type: QUERY -PREHOOK: Input: default@a_n0 -PREHOOK: Input: default@b_n0 -PREHOOK: Input: default@c_n0 -PREHOOK: Input: default@d_n0 -PREHOOK: Input: default@e_n0 -#### A masked pattern was here #### -POSTHOOK: query: explain select * from a_n0 join b_n0 on a_n0.val1=b_n0.val1 join c_n0 on a_n0.val1=c_n0.val1 join d_n0 on a_n0.val1=d_n0.val1 join e_n0 on a_n0.val2=e_n0.val2 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@a_n0 -POSTHOOK: Input: default@b_n0 -POSTHOOK: Input: default@c_n0 -POSTHOOK: Input: default@d_n0 -POSTHOOK: Input: default@e_n0 -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-3 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-3 - Stage-0 depends on stages: Stage-4 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: a_n0 - filterExpr: (val1 is not null and val2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: (val1 is not null and val2 is not null) (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val1 (type: int), val2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - TableScan - alias: b_n0 - filterExpr: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val1 (type: int), val2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int) - TableScan - alias: c_n0 - filterExpr: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val1 (type: int), val2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int) - TableScan - alias: d_n0 - filterExpr: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val1 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val1 (type: int), val2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: int) - 1 _col0 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-4 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int) - TableScan - alias: e_n0 - filterExpr: val2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Filter Operator - predicate: val2 is not null (type: boolean) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: val1 (type: int), val2 (type: int) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col1 (type: int) - null sort order: z - sort order: + - Map-reduce partition columns: _col1 (type: int) - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - value expressions: _col0 (type: int) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col1 (type: int) - 1 _col1 (type: int) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8, _col9 - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1 Data size: 8 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink - -PREHOOK: query: explain select * from src a_n0 join src b_n0 on a_n0.key=b_n0.key left outer join src c_n0 on b_n0.key=c_n0.key and b_n0.key<10 -PREHOOK: type: QUERY -PREHOOK: Input: default@src -#### A masked pattern was here #### -POSTHOOK: query: explain select * from src a_n0 join src b_n0 on a_n0.key=b_n0.key left outer join src c_n0 on b_n0.key=c_n0.key and b_n0.key<10 -POSTHOOK: type: QUERY -POSTHOOK: Input: default@src -#### A masked pattern was here #### -STAGE DEPENDENCIES: - Stage-1 is a root stage - Stage-2 depends on stages: Stage-1 - Stage-0 depends on stages: Stage-2 - -STAGE PLANS: - Stage: Stage-1 - Map Reduce - Map Operator Tree: - TableScan - alias: b_n0 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string), (UDFToDouble(key) < 10.0D) (type: boolean) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean) - TableScan - alias: a_n0 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Inner Join 0 to 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 791 Data size: 284760 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - - Stage: Stage-2 - Map Reduce - Map Operator Tree: - TableScan - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 791 Data size: 284760 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string), _col2 (type: boolean), _col3 (type: string), _col4 (type: string) - TableScan - alias: c_n0 - filterExpr: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Filter Operator - predicate: key is not null (type: boolean) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: key (type: string), value (type: string) - outputColumnNames: _col0, _col1 - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - Reduce Output Operator - key expressions: _col0 (type: string) - null sort order: z - sort order: + - Map-reduce partition columns: _col0 (type: string) - Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - value expressions: _col1 (type: string) - Reduce Operator Tree: - Join Operator - condition map: - Left Outer Join 0 to 1 - filter predicates: - 0 {VALUE._col1} - 1 - keys: - 0 _col0 (type: string) - 1 _col0 (type: string) - outputColumnNames: _col0, _col1, _col3, _col4, _col5, _col6 - Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE - Select Operator - expressions: _col3 (type: string), _col4 (type: string), _col0 (type: string), _col1 (type: string), _col5 (type: string), _col6 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5 - Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false - Statistics: Num rows: 1251 Data size: 668034 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-0 - Fetch Operator - limit: -1 - Processor Tree: - ListSink -